May 15 15:12:46.861828 kernel: Linux version 6.12.20-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Thu May 15 10:42:41 -00 2025 May 15 15:12:46.861855 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=48287e633374b880fa618bd42bee102ae77c50831859c6cedd6ca9e1aec3dd5c May 15 15:12:46.861865 kernel: BIOS-provided physical RAM map: May 15 15:12:46.861872 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable May 15 15:12:46.861878 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved May 15 15:12:46.861885 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved May 15 15:12:46.861893 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable May 15 15:12:46.861903 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved May 15 15:12:46.861913 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 15 15:12:46.861920 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved May 15 15:12:46.861927 kernel: NX (Execute Disable) protection: active May 15 15:12:46.861933 kernel: APIC: Static calls initialized May 15 15:12:46.861940 kernel: SMBIOS 2.8 present. May 15 15:12:46.861947 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 May 15 15:12:46.861970 kernel: DMI: Memory slots populated: 1/1 May 15 15:12:46.861978 kernel: Hypervisor detected: KVM May 15 15:12:46.862057 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 15 15:12:46.862069 kernel: kvm-clock: using sched offset of 4339864704 cycles May 15 15:12:46.862083 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 15 15:12:46.862095 kernel: tsc: Detected 2494.136 MHz processor May 15 15:12:46.862103 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 15 15:12:46.862112 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 15 15:12:46.862120 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 May 15 15:12:46.862132 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs May 15 15:12:46.862140 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 15 15:12:46.862148 kernel: ACPI: Early table checksum verification disabled May 15 15:12:46.862155 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) May 15 15:12:46.862164 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 15:12:46.862172 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 15:12:46.862180 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 15:12:46.862188 kernel: ACPI: FACS 0x000000007FFE0000 000040 May 15 15:12:46.862195 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 15:12:46.862206 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 15:12:46.862214 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 15:12:46.862222 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 15:12:46.862230 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] May 15 15:12:46.862238 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] May 15 15:12:46.862246 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] May 15 15:12:46.862253 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] May 15 15:12:46.862262 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] May 15 15:12:46.862276 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] May 15 15:12:46.862285 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] May 15 15:12:46.862293 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] May 15 15:12:46.862301 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] May 15 15:12:46.862310 kernel: NUMA: Node 0 [mem 0x00001000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00001000-0x7ffdafff] May 15 15:12:46.862319 kernel: NODE_DATA(0) allocated [mem 0x7ffd3dc0-0x7ffdafff] May 15 15:12:46.862330 kernel: Zone ranges: May 15 15:12:46.862339 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 15 15:12:46.862347 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] May 15 15:12:46.862355 kernel: Normal empty May 15 15:12:46.862364 kernel: Device empty May 15 15:12:46.862372 kernel: Movable zone start for each node May 15 15:12:46.862381 kernel: Early memory node ranges May 15 15:12:46.862389 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] May 15 15:12:46.862398 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] May 15 15:12:46.862409 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] May 15 15:12:46.862417 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 15 15:12:46.862425 kernel: On node 0, zone DMA: 97 pages in unavailable ranges May 15 15:12:46.862434 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges May 15 15:12:46.862442 kernel: ACPI: PM-Timer IO Port: 0x608 May 15 15:12:46.862450 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 15 15:12:46.862462 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 15 15:12:46.862470 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) May 15 15:12:46.862480 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 15 15:12:46.862491 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 15 15:12:46.862501 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 15 15:12:46.862510 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 15 15:12:46.862518 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 15 15:12:46.862527 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 15 15:12:46.862535 kernel: TSC deadline timer available May 15 15:12:46.862544 kernel: CPU topo: Max. logical packages: 1 May 15 15:12:46.862552 kernel: CPU topo: Max. logical dies: 1 May 15 15:12:46.862560 kernel: CPU topo: Max. dies per package: 1 May 15 15:12:46.862568 kernel: CPU topo: Max. threads per core: 1 May 15 15:12:46.862580 kernel: CPU topo: Num. cores per package: 2 May 15 15:12:46.862588 kernel: CPU topo: Num. threads per package: 2 May 15 15:12:46.862597 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs May 15 15:12:46.862606 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() May 15 15:12:46.862614 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices May 15 15:12:46.862623 kernel: Booting paravirtualized kernel on KVM May 15 15:12:46.862631 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 15 15:12:46.862639 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 May 15 15:12:46.862648 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 May 15 15:12:46.862660 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 May 15 15:12:46.862668 kernel: pcpu-alloc: [0] 0 1 May 15 15:12:46.862683 kernel: kvm-guest: PV spinlocks disabled, no host support May 15 15:12:46.862698 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=48287e633374b880fa618bd42bee102ae77c50831859c6cedd6ca9e1aec3dd5c May 15 15:12:46.862710 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 15 15:12:46.862722 kernel: random: crng init done May 15 15:12:46.862733 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 15 15:12:46.862745 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) May 15 15:12:46.862761 kernel: Fallback order for Node 0: 0 May 15 15:12:46.862774 kernel: Built 1 zonelists, mobility grouping on. Total pages: 524153 May 15 15:12:46.862786 kernel: Policy zone: DMA32 May 15 15:12:46.862798 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 15 15:12:46.862811 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 15 15:12:46.862820 kernel: Kernel/User page tables isolation: enabled May 15 15:12:46.862829 kernel: ftrace: allocating 40065 entries in 157 pages May 15 15:12:46.862838 kernel: ftrace: allocated 157 pages with 5 groups May 15 15:12:46.862846 kernel: Dynamic Preempt: voluntary May 15 15:12:46.862858 kernel: rcu: Preemptible hierarchical RCU implementation. May 15 15:12:46.862872 kernel: rcu: RCU event tracing is enabled. May 15 15:12:46.862882 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 15 15:12:46.862890 kernel: Trampoline variant of Tasks RCU enabled. May 15 15:12:46.862899 kernel: Rude variant of Tasks RCU enabled. May 15 15:12:46.862907 kernel: Tracing variant of Tasks RCU enabled. May 15 15:12:46.862915 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 15 15:12:46.862924 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 15 15:12:46.862932 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 15 15:12:46.862947 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 15 15:12:46.864027 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 15 15:12:46.864051 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 May 15 15:12:46.864060 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 15 15:12:46.864069 kernel: Console: colour VGA+ 80x25 May 15 15:12:46.864078 kernel: printk: legacy console [tty0] enabled May 15 15:12:46.864086 kernel: printk: legacy console [ttyS0] enabled May 15 15:12:46.864094 kernel: ACPI: Core revision 20240827 May 15 15:12:46.864103 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns May 15 15:12:46.864126 kernel: APIC: Switch to symmetric I/O mode setup May 15 15:12:46.864135 kernel: x2apic enabled May 15 15:12:46.864144 kernel: APIC: Switched APIC routing to: physical x2apic May 15 15:12:46.864156 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 15 15:12:46.864169 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39654230, max_idle_ns: 440795207432 ns May 15 15:12:46.864179 kernel: Calibrating delay loop (skipped) preset value.. 4988.27 BogoMIPS (lpj=2494136) May 15 15:12:46.864190 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 May 15 15:12:46.864205 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 May 15 15:12:46.864218 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 15 15:12:46.864235 kernel: Spectre V2 : Mitigation: Retpolines May 15 15:12:46.864247 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch May 15 15:12:46.864261 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT May 15 15:12:46.864273 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls May 15 15:12:46.864287 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 15 15:12:46.864297 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl May 15 15:12:46.864307 kernel: MDS: Mitigation: Clear CPU buffers May 15 15:12:46.864318 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode May 15 15:12:46.864327 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 15 15:12:46.864336 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 15 15:12:46.864345 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 15 15:12:46.864354 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 15 15:12:46.864363 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. May 15 15:12:46.864372 kernel: Freeing SMP alternatives memory: 32K May 15 15:12:46.864381 kernel: pid_max: default: 32768 minimum: 301 May 15 15:12:46.864390 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima May 15 15:12:46.864402 kernel: landlock: Up and running. May 15 15:12:46.864411 kernel: SELinux: Initializing. May 15 15:12:46.864419 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 15 15:12:46.864428 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 15 15:12:46.864437 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) May 15 15:12:46.864446 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. May 15 15:12:46.864455 kernel: signal: max sigframe size: 1776 May 15 15:12:46.864464 kernel: rcu: Hierarchical SRCU implementation. May 15 15:12:46.864474 kernel: rcu: Max phase no-delay instances is 400. May 15 15:12:46.864485 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level May 15 15:12:46.864494 kernel: NMI watchdog: Perf NMI watchdog permanently disabled May 15 15:12:46.864503 kernel: smp: Bringing up secondary CPUs ... May 15 15:12:46.864512 kernel: smpboot: x86: Booting SMP configuration: May 15 15:12:46.864525 kernel: .... node #0, CPUs: #1 May 15 15:12:46.864534 kernel: smp: Brought up 1 node, 2 CPUs May 15 15:12:46.864543 kernel: smpboot: Total of 2 processors activated (9976.54 BogoMIPS) May 15 15:12:46.864552 kernel: Memory: 1966908K/2096612K available (14336K kernel code, 2438K rwdata, 9944K rodata, 54416K init, 2544K bss, 125140K reserved, 0K cma-reserved) May 15 15:12:46.864561 kernel: devtmpfs: initialized May 15 15:12:46.864573 kernel: x86/mm: Memory block size: 128MB May 15 15:12:46.864582 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 15 15:12:46.864591 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 15 15:12:46.864600 kernel: pinctrl core: initialized pinctrl subsystem May 15 15:12:46.864610 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 15 15:12:46.864618 kernel: audit: initializing netlink subsys (disabled) May 15 15:12:46.864627 kernel: audit: type=2000 audit(1747321962.561:1): state=initialized audit_enabled=0 res=1 May 15 15:12:46.864641 kernel: thermal_sys: Registered thermal governor 'step_wise' May 15 15:12:46.864657 kernel: thermal_sys: Registered thermal governor 'user_space' May 15 15:12:46.864671 kernel: cpuidle: using governor menu May 15 15:12:46.864680 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 15 15:12:46.864696 kernel: dca service started, version 1.12.1 May 15 15:12:46.864706 kernel: PCI: Using configuration type 1 for base access May 15 15:12:46.864716 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 15 15:12:46.864725 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 15 15:12:46.864734 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page May 15 15:12:46.864742 kernel: ACPI: Added _OSI(Module Device) May 15 15:12:46.864751 kernel: ACPI: Added _OSI(Processor Device) May 15 15:12:46.864764 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 15 15:12:46.864773 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 15 15:12:46.864782 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 15 15:12:46.864791 kernel: ACPI: Interpreter enabled May 15 15:12:46.864799 kernel: ACPI: PM: (supports S0 S5) May 15 15:12:46.864808 kernel: ACPI: Using IOAPIC for interrupt routing May 15 15:12:46.864817 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 15 15:12:46.864826 kernel: PCI: Using E820 reservations for host bridge windows May 15 15:12:46.864835 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F May 15 15:12:46.864847 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 15 15:12:46.865078 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] May 15 15:12:46.865176 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] May 15 15:12:46.865264 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge May 15 15:12:46.865276 kernel: acpiphp: Slot [3] registered May 15 15:12:46.865285 kernel: acpiphp: Slot [4] registered May 15 15:12:46.865294 kernel: acpiphp: Slot [5] registered May 15 15:12:46.865303 kernel: acpiphp: Slot [6] registered May 15 15:12:46.865317 kernel: acpiphp: Slot [7] registered May 15 15:12:46.865325 kernel: acpiphp: Slot [8] registered May 15 15:12:46.865334 kernel: acpiphp: Slot [9] registered May 15 15:12:46.865343 kernel: acpiphp: Slot [10] registered May 15 15:12:46.865351 kernel: acpiphp: Slot [11] registered May 15 15:12:46.865360 kernel: acpiphp: Slot [12] registered May 15 15:12:46.865369 kernel: acpiphp: Slot [13] registered May 15 15:12:46.865377 kernel: acpiphp: Slot [14] registered May 15 15:12:46.865386 kernel: acpiphp: Slot [15] registered May 15 15:12:46.865398 kernel: acpiphp: Slot [16] registered May 15 15:12:46.865407 kernel: acpiphp: Slot [17] registered May 15 15:12:46.865415 kernel: acpiphp: Slot [18] registered May 15 15:12:46.865424 kernel: acpiphp: Slot [19] registered May 15 15:12:46.865433 kernel: acpiphp: Slot [20] registered May 15 15:12:46.865441 kernel: acpiphp: Slot [21] registered May 15 15:12:46.865450 kernel: acpiphp: Slot [22] registered May 15 15:12:46.865459 kernel: acpiphp: Slot [23] registered May 15 15:12:46.865468 kernel: acpiphp: Slot [24] registered May 15 15:12:46.865477 kernel: acpiphp: Slot [25] registered May 15 15:12:46.865489 kernel: acpiphp: Slot [26] registered May 15 15:12:46.865497 kernel: acpiphp: Slot [27] registered May 15 15:12:46.865506 kernel: acpiphp: Slot [28] registered May 15 15:12:46.865515 kernel: acpiphp: Slot [29] registered May 15 15:12:46.865524 kernel: acpiphp: Slot [30] registered May 15 15:12:46.865533 kernel: acpiphp: Slot [31] registered May 15 15:12:46.865542 kernel: PCI host bridge to bus 0000:00 May 15 15:12:46.865708 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 15 15:12:46.865978 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 15 15:12:46.866098 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 15 15:12:46.866182 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] May 15 15:12:46.866265 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] May 15 15:12:46.866347 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 15 15:12:46.866476 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint May 15 15:12:46.866596 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint May 15 15:12:46.866707 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 conventional PCI endpoint May 15 15:12:46.866800 kernel: pci 0000:00:01.1: BAR 4 [io 0xc1e0-0xc1ef] May 15 15:12:46.866891 kernel: pci 0000:00:01.1: BAR 0 [io 0x01f0-0x01f7]: legacy IDE quirk May 15 15:12:46.866997 kernel: pci 0000:00:01.1: BAR 1 [io 0x03f6]: legacy IDE quirk May 15 15:12:46.867088 kernel: pci 0000:00:01.1: BAR 2 [io 0x0170-0x0177]: legacy IDE quirk May 15 15:12:46.867181 kernel: pci 0000:00:01.1: BAR 3 [io 0x0376]: legacy IDE quirk May 15 15:12:46.867290 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 conventional PCI endpoint May 15 15:12:46.867404 kernel: pci 0000:00:01.2: BAR 4 [io 0xc180-0xc19f] May 15 15:12:46.867508 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 conventional PCI endpoint May 15 15:12:46.867600 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI May 15 15:12:46.867691 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB May 15 15:12:46.867800 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 conventional PCI endpoint May 15 15:12:46.867894 kernel: pci 0000:00:02.0: BAR 0 [mem 0xfe000000-0xfe7fffff pref] May 15 15:12:46.868022 kernel: pci 0000:00:02.0: BAR 2 [mem 0xfe800000-0xfe803fff 64bit pref] May 15 15:12:46.868193 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfebf0000-0xfebf0fff] May 15 15:12:46.868319 kernel: pci 0000:00:02.0: ROM [mem 0xfebe0000-0xfebeffff pref] May 15 15:12:46.868442 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 15 15:12:46.868566 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint May 15 15:12:46.868660 kernel: pci 0000:00:03.0: BAR 0 [io 0xc1a0-0xc1bf] May 15 15:12:46.868756 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebf1000-0xfebf1fff] May 15 15:12:46.869046 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe804000-0xfe807fff 64bit pref] May 15 15:12:46.869157 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint May 15 15:12:46.869260 kernel: pci 0000:00:04.0: BAR 0 [io 0xc1c0-0xc1df] May 15 15:12:46.869368 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebf2000-0xfebf2fff] May 15 15:12:46.869457 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe808000-0xfe80bfff 64bit pref] May 15 15:12:46.869569 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 conventional PCI endpoint May 15 15:12:46.869665 kernel: pci 0000:00:05.0: BAR 0 [io 0xc100-0xc13f] May 15 15:12:46.869754 kernel: pci 0000:00:05.0: BAR 1 [mem 0xfebf3000-0xfebf3fff] May 15 15:12:46.869842 kernel: pci 0000:00:05.0: BAR 4 [mem 0xfe80c000-0xfe80ffff 64bit pref] May 15 15:12:46.869953 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint May 15 15:12:46.870111 kernel: pci 0000:00:06.0: BAR 0 [io 0xc000-0xc07f] May 15 15:12:46.870202 kernel: pci 0000:00:06.0: BAR 1 [mem 0xfebf4000-0xfebf4fff] May 15 15:12:46.870292 kernel: pci 0000:00:06.0: BAR 4 [mem 0xfe810000-0xfe813fff 64bit pref] May 15 15:12:46.870393 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint May 15 15:12:46.870483 kernel: pci 0000:00:07.0: BAR 0 [io 0xc080-0xc0ff] May 15 15:12:46.870598 kernel: pci 0000:00:07.0: BAR 1 [mem 0xfebf5000-0xfebf5fff] May 15 15:12:46.870688 kernel: pci 0000:00:07.0: BAR 4 [mem 0xfe814000-0xfe817fff 64bit pref] May 15 15:12:46.870789 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 conventional PCI endpoint May 15 15:12:46.870882 kernel: pci 0000:00:08.0: BAR 0 [io 0xc140-0xc17f] May 15 15:12:46.871625 kernel: pci 0000:00:08.0: BAR 4 [mem 0xfe818000-0xfe81bfff 64bit pref] May 15 15:12:46.871646 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 15 15:12:46.871657 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 15 15:12:46.871666 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 15 15:12:46.871675 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 15 15:12:46.871684 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 May 15 15:12:46.871694 kernel: iommu: Default domain type: Translated May 15 15:12:46.871703 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 15 15:12:46.871712 kernel: PCI: Using ACPI for IRQ routing May 15 15:12:46.871726 kernel: PCI: pci_cache_line_size set to 64 bytes May 15 15:12:46.871735 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] May 15 15:12:46.871744 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] May 15 15:12:46.871853 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device May 15 15:12:46.871946 kernel: pci 0000:00:02.0: vgaarb: bridge control possible May 15 15:12:46.872053 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 15 15:12:46.872065 kernel: vgaarb: loaded May 15 15:12:46.872074 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 May 15 15:12:46.872087 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter May 15 15:12:46.872096 kernel: clocksource: Switched to clocksource kvm-clock May 15 15:12:46.872114 kernel: VFS: Disk quotas dquot_6.6.0 May 15 15:12:46.872123 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 15 15:12:46.872132 kernel: pnp: PnP ACPI init May 15 15:12:46.872141 kernel: pnp: PnP ACPI: found 4 devices May 15 15:12:46.872150 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 15 15:12:46.872159 kernel: NET: Registered PF_INET protocol family May 15 15:12:46.872168 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) May 15 15:12:46.872182 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) May 15 15:12:46.872191 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 15 15:12:46.872199 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) May 15 15:12:46.872208 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) May 15 15:12:46.872217 kernel: TCP: Hash tables configured (established 16384 bind 16384) May 15 15:12:46.872226 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) May 15 15:12:46.872235 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) May 15 15:12:46.872244 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 15 15:12:46.872253 kernel: NET: Registered PF_XDP protocol family May 15 15:12:46.872345 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 15 15:12:46.872425 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 15 15:12:46.872504 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 15 15:12:46.872584 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] May 15 15:12:46.872696 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] May 15 15:12:46.872791 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release May 15 15:12:46.872924 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 15 15:12:46.872943 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 May 15 15:12:46.873417 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x720 took 24422 usecs May 15 15:12:46.873439 kernel: PCI: CLS 0 bytes, default 64 May 15 15:12:46.873452 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer May 15 15:12:46.873464 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39654230, max_idle_ns: 440795207432 ns May 15 15:12:46.873476 kernel: Initialise system trusted keyrings May 15 15:12:46.873489 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 May 15 15:12:46.873503 kernel: Key type asymmetric registered May 15 15:12:46.873515 kernel: Asymmetric key parser 'x509' registered May 15 15:12:46.873588 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) May 15 15:12:46.873603 kernel: io scheduler mq-deadline registered May 15 15:12:46.873612 kernel: io scheduler kyber registered May 15 15:12:46.873622 kernel: io scheduler bfq registered May 15 15:12:46.873631 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 15 15:12:46.873640 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 May 15 15:12:46.873650 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 May 15 15:12:46.873659 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 May 15 15:12:46.873668 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 15 15:12:46.873677 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 15 15:12:46.873689 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 15 15:12:46.873698 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 15 15:12:46.873707 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 15 15:12:46.873717 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 15 15:12:46.873846 kernel: rtc_cmos 00:03: RTC can wake from S4 May 15 15:12:46.873936 kernel: rtc_cmos 00:03: registered as rtc0 May 15 15:12:46.874061 kernel: rtc_cmos 00:03: setting system clock to 2025-05-15T15:12:46 UTC (1747321966) May 15 15:12:46.874152 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram May 15 15:12:46.874164 kernel: intel_pstate: CPU model not supported May 15 15:12:46.874173 kernel: NET: Registered PF_INET6 protocol family May 15 15:12:46.874182 kernel: Segment Routing with IPv6 May 15 15:12:46.874192 kernel: In-situ OAM (IOAM) with IPv6 May 15 15:12:46.874201 kernel: NET: Registered PF_PACKET protocol family May 15 15:12:46.874210 kernel: Key type dns_resolver registered May 15 15:12:46.874219 kernel: IPI shorthand broadcast: enabled May 15 15:12:46.874228 kernel: sched_clock: Marking stable (4335005344, 90246640)->(4439887964, -14635980) May 15 15:12:46.874237 kernel: registered taskstats version 1 May 15 15:12:46.874249 kernel: Loading compiled-in X.509 certificates May 15 15:12:46.874258 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.20-flatcar: 05e05785144663be6df1db78301487421c4773b6' May 15 15:12:46.874267 kernel: Demotion targets for Node 0: null May 15 15:12:46.874276 kernel: Key type .fscrypt registered May 15 15:12:46.874285 kernel: Key type fscrypt-provisioning registered May 15 15:12:46.874313 kernel: ima: No TPM chip found, activating TPM-bypass! May 15 15:12:46.874325 kernel: ima: Allocated hash algorithm: sha1 May 15 15:12:46.874335 kernel: ima: No architecture policies found May 15 15:12:46.874347 kernel: clk: Disabling unused clocks May 15 15:12:46.874356 kernel: Warning: unable to open an initial console. May 15 15:12:46.874365 kernel: Freeing unused kernel image (initmem) memory: 54416K May 15 15:12:46.874375 kernel: Write protecting the kernel read-only data: 24576k May 15 15:12:46.874384 kernel: Freeing unused kernel image (rodata/data gap) memory: 296K May 15 15:12:46.874396 kernel: Run /init as init process May 15 15:12:46.874406 kernel: with arguments: May 15 15:12:46.874415 kernel: /init May 15 15:12:46.874424 kernel: with environment: May 15 15:12:46.874436 kernel: HOME=/ May 15 15:12:46.874445 kernel: TERM=linux May 15 15:12:46.874454 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 15 15:12:46.874465 systemd[1]: Successfully made /usr/ read-only. May 15 15:12:46.874478 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 15 15:12:46.874488 systemd[1]: Detected virtualization kvm. May 15 15:12:46.874498 systemd[1]: Detected architecture x86-64. May 15 15:12:46.874507 systemd[1]: Running in initrd. May 15 15:12:46.874519 systemd[1]: No hostname configured, using default hostname. May 15 15:12:46.874529 systemd[1]: Hostname set to . May 15 15:12:46.874539 systemd[1]: Initializing machine ID from VM UUID. May 15 15:12:46.874548 systemd[1]: Queued start job for default target initrd.target. May 15 15:12:46.874558 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 15 15:12:46.874567 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 15 15:12:46.874578 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 15 15:12:46.874587 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 15 15:12:46.874600 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 15 15:12:46.874613 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 15 15:12:46.874624 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 15 15:12:46.874636 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 15 15:12:46.874646 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 15 15:12:46.874656 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 15 15:12:46.874665 systemd[1]: Reached target paths.target - Path Units. May 15 15:12:46.874675 systemd[1]: Reached target slices.target - Slice Units. May 15 15:12:46.874685 systemd[1]: Reached target swap.target - Swaps. May 15 15:12:46.874695 systemd[1]: Reached target timers.target - Timer Units. May 15 15:12:46.874704 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 15 15:12:46.874714 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 15 15:12:46.874727 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 15 15:12:46.874737 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 15 15:12:46.874746 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 15 15:12:46.874756 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 15 15:12:46.874766 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 15 15:12:46.874776 systemd[1]: Reached target sockets.target - Socket Units. May 15 15:12:46.874786 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 15 15:12:46.874795 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 15 15:12:46.874808 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 15 15:12:46.874818 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). May 15 15:12:46.874828 systemd[1]: Starting systemd-fsck-usr.service... May 15 15:12:46.874838 systemd[1]: Starting systemd-journald.service - Journal Service... May 15 15:12:46.874848 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 15 15:12:46.874857 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 15 15:12:46.874867 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 15 15:12:46.874881 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 15 15:12:46.874891 systemd[1]: Finished systemd-fsck-usr.service. May 15 15:12:46.874901 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 15 15:12:46.874938 systemd-journald[211]: Collecting audit messages is disabled. May 15 15:12:46.875001 systemd-journald[211]: Journal started May 15 15:12:46.875024 systemd-journald[211]: Runtime Journal (/run/log/journal/782f0142214d40feb3e3035039648a79) is 4.9M, max 39.5M, 34.6M free. May 15 15:12:46.876986 systemd[1]: Started systemd-journald.service - Journal Service. May 15 15:12:46.868119 systemd-modules-load[212]: Inserted module 'overlay' May 15 15:12:46.879098 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 15 15:12:46.885622 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 15 15:12:46.920109 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 15 15:12:46.920139 kernel: Bridge firewalling registered May 15 15:12:46.902849 systemd-tmpfiles[223]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. May 15 15:12:46.906368 systemd-modules-load[212]: Inserted module 'br_netfilter' May 15 15:12:46.921398 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 15 15:12:46.922332 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 15 15:12:46.923331 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 15 15:12:46.926242 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 15 15:12:46.927324 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 15 15:12:46.930170 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 15 15:12:46.953386 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 15 15:12:46.956027 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 15 15:12:46.960176 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 15 15:12:46.963014 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 15 15:12:46.964764 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 15 15:12:46.990757 dracut-cmdline[250]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=48287e633374b880fa618bd42bee102ae77c50831859c6cedd6ca9e1aec3dd5c May 15 15:12:47.008085 systemd-resolved[247]: Positive Trust Anchors: May 15 15:12:47.008097 systemd-resolved[247]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 15 15:12:47.008133 systemd-resolved[247]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 15 15:12:47.013555 systemd-resolved[247]: Defaulting to hostname 'linux'. May 15 15:12:47.015673 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 15 15:12:47.016086 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 15 15:12:47.094060 kernel: SCSI subsystem initialized May 15 15:12:47.107024 kernel: Loading iSCSI transport class v2.0-870. May 15 15:12:47.124007 kernel: iscsi: registered transport (tcp) May 15 15:12:47.154050 kernel: iscsi: registered transport (qla4xxx) May 15 15:12:47.154138 kernel: QLogic iSCSI HBA Driver May 15 15:12:47.178339 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 15 15:12:47.198237 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 15 15:12:47.200673 systemd[1]: Reached target network-pre.target - Preparation for Network. May 15 15:12:47.255154 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 15 15:12:47.257539 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 15 15:12:47.316006 kernel: raid6: avx2x4 gen() 15492 MB/s May 15 15:12:47.333008 kernel: raid6: avx2x2 gen() 16790 MB/s May 15 15:12:47.350227 kernel: raid6: avx2x1 gen() 13078 MB/s May 15 15:12:47.350301 kernel: raid6: using algorithm avx2x2 gen() 16790 MB/s May 15 15:12:47.368132 kernel: raid6: .... xor() 19392 MB/s, rmw enabled May 15 15:12:47.368224 kernel: raid6: using avx2x2 recovery algorithm May 15 15:12:47.391021 kernel: xor: automatically using best checksumming function avx May 15 15:12:47.576042 kernel: Btrfs loaded, zoned=no, fsverity=no May 15 15:12:47.584617 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 15 15:12:47.586768 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 15 15:12:47.616840 systemd-udevd[460]: Using default interface naming scheme 'v255'. May 15 15:12:47.623746 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 15 15:12:47.628820 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 15 15:12:47.653876 dracut-pre-trigger[468]: rd.md=0: removing MD RAID activation May 15 15:12:47.684881 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 15 15:12:47.686286 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 15 15:12:47.759881 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 15 15:12:47.762007 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 15 15:12:47.847002 kernel: virtio_scsi virtio3: 2/0/0 default/read/poll queues May 15 15:12:47.926980 kernel: cryptd: max_cpu_qlen set to 1000 May 15 15:12:47.927008 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues May 15 15:12:47.950167 kernel: scsi host0: Virtio SCSI HBA May 15 15:12:47.950304 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) May 15 15:12:47.950409 kernel: libata version 3.00 loaded. May 15 15:12:47.950423 kernel: AES CTR mode by8 optimization enabled May 15 15:12:47.950443 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 May 15 15:12:47.950456 kernel: ACPI: bus type USB registered May 15 15:12:47.950467 kernel: usbcore: registered new interface driver usbfs May 15 15:12:47.950479 kernel: usbcore: registered new interface driver hub May 15 15:12:47.950491 kernel: usbcore: registered new device driver usb May 15 15:12:47.950503 kernel: ata_piix 0000:00:01.1: version 2.13 May 15 15:12:47.966634 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 15 15:12:47.966654 kernel: GPT:9289727 != 125829119 May 15 15:12:47.966667 kernel: GPT:Alternate GPT header not at the end of the disk. May 15 15:12:47.966685 kernel: GPT:9289727 != 125829119 May 15 15:12:47.966696 kernel: GPT: Use GNU Parted to correct GPT errors. May 15 15:12:47.966708 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 15 15:12:47.966721 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues May 15 15:12:47.966853 kernel: scsi host1: ata_piix May 15 15:12:47.966999 kernel: virtio_blk virtio5: [vdb] 980 512-byte logical blocks (502 kB/490 KiB) May 15 15:12:47.967151 kernel: scsi host2: ata_piix May 15 15:12:47.967371 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 lpm-pol 0 May 15 15:12:47.967388 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 lpm-pol 0 May 15 15:12:47.938289 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 15 15:12:47.938423 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 15 15:12:47.939032 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 15 15:12:47.943254 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 15 15:12:47.954761 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 15 15:12:47.981218 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller May 15 15:12:47.989662 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 May 15 15:12:47.989863 kernel: uhci_hcd 0000:00:01.2: detected 2 ports May 15 15:12:47.990102 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 May 15 15:12:47.990265 kernel: hub 1-0:1.0: USB hub found May 15 15:12:47.990470 kernel: hub 1-0:1.0: 2 ports detected May 15 15:12:48.018472 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 15 15:12:48.175130 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. May 15 15:12:48.176034 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 15 15:12:48.186476 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. May 15 15:12:48.198319 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 15 15:12:48.206176 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. May 15 15:12:48.206637 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. May 15 15:12:48.207762 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 15 15:12:48.208550 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 15 15:12:48.209479 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 15 15:12:48.211303 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 15 15:12:48.212611 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 15 15:12:48.232028 disk-uuid[617]: Primary Header is updated. May 15 15:12:48.232028 disk-uuid[617]: Secondary Entries is updated. May 15 15:12:48.232028 disk-uuid[617]: Secondary Header is updated. May 15 15:12:48.240748 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 15 15:12:48.245579 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 15 15:12:48.255019 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 15 15:12:49.251099 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 15 15:12:49.251431 disk-uuid[620]: The operation has completed successfully. May 15 15:12:49.317796 systemd[1]: disk-uuid.service: Deactivated successfully. May 15 15:12:49.318002 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 15 15:12:49.359312 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 15 15:12:49.381054 sh[636]: Success May 15 15:12:49.401998 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 15 15:12:49.402116 kernel: device-mapper: uevent: version 1.0.3 May 15 15:12:49.402130 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev May 15 15:12:49.416013 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" May 15 15:12:49.487641 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 15 15:12:49.489023 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 15 15:12:49.508948 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 15 15:12:49.522576 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' May 15 15:12:49.522686 kernel: BTRFS: device fsid 2d504097-db49-4d66-a0d5-eeb665b21004 devid 1 transid 41 /dev/mapper/usr (253:0) scanned by mount (648) May 15 15:12:49.523137 kernel: BTRFS info (device dm-0): first mount of filesystem 2d504097-db49-4d66-a0d5-eeb665b21004 May 15 15:12:49.525157 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm May 15 15:12:49.525218 kernel: BTRFS info (device dm-0): using free-space-tree May 15 15:12:49.535853 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 15 15:12:49.537154 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. May 15 15:12:49.537874 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 15 15:12:49.539220 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 15 15:12:49.540772 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 15 15:12:49.567848 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (677) May 15 15:12:49.567916 kernel: BTRFS info (device vda6): first mount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 15:12:49.569674 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 15 15:12:49.569743 kernel: BTRFS info (device vda6): using free-space-tree May 15 15:12:49.581055 kernel: BTRFS info (device vda6): last unmount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 15:12:49.582855 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 15 15:12:49.584782 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 15 15:12:49.695397 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 15 15:12:49.700172 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 15 15:12:49.756679 systemd-networkd[820]: lo: Link UP May 15 15:12:49.756690 systemd-networkd[820]: lo: Gained carrier May 15 15:12:49.759399 systemd-networkd[820]: Enumeration completed May 15 15:12:49.759734 systemd-networkd[820]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. May 15 15:12:49.759738 systemd-networkd[820]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. May 15 15:12:49.759902 systemd[1]: Started systemd-networkd.service - Network Configuration. May 15 15:12:49.760535 systemd[1]: Reached target network.target - Network. May 15 15:12:49.761557 systemd-networkd[820]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 15 15:12:49.761562 systemd-networkd[820]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. May 15 15:12:49.762880 systemd-networkd[820]: eth0: Link UP May 15 15:12:49.762885 systemd-networkd[820]: eth0: Gained carrier May 15 15:12:49.762899 systemd-networkd[820]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. May 15 15:12:49.772433 systemd-networkd[820]: eth1: Link UP May 15 15:12:49.772444 systemd-networkd[820]: eth1: Gained carrier May 15 15:12:49.772462 systemd-networkd[820]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 15 15:12:49.792050 systemd-networkd[820]: eth0: DHCPv4 address 24.144.90.91/20, gateway 24.144.80.1 acquired from 169.254.169.253 May 15 15:12:49.803794 ignition[722]: Ignition 2.21.0 May 15 15:12:49.803809 ignition[722]: Stage: fetch-offline May 15 15:12:49.803863 ignition[722]: no configs at "/usr/lib/ignition/base.d" May 15 15:12:49.803873 ignition[722]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 15:12:49.803985 ignition[722]: parsed url from cmdline: "" May 15 15:12:49.807373 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 15 15:12:49.803989 ignition[722]: no config URL provided May 15 15:12:49.808046 systemd-networkd[820]: eth1: DHCPv4 address 10.124.0.32/20 acquired from 169.254.169.253 May 15 15:12:49.803994 ignition[722]: reading system config file "/usr/lib/ignition/user.ign" May 15 15:12:49.804002 ignition[722]: no config at "/usr/lib/ignition/user.ign" May 15 15:12:49.811102 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... May 15 15:12:49.804007 ignition[722]: failed to fetch config: resource requires networking May 15 15:12:49.804185 ignition[722]: Ignition finished successfully May 15 15:12:49.852150 ignition[829]: Ignition 2.21.0 May 15 15:12:49.852166 ignition[829]: Stage: fetch May 15 15:12:49.852667 ignition[829]: no configs at "/usr/lib/ignition/base.d" May 15 15:12:49.852687 ignition[829]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 15:12:49.852818 ignition[829]: parsed url from cmdline: "" May 15 15:12:49.852825 ignition[829]: no config URL provided May 15 15:12:49.852832 ignition[829]: reading system config file "/usr/lib/ignition/user.ign" May 15 15:12:49.852844 ignition[829]: no config at "/usr/lib/ignition/user.ign" May 15 15:12:49.852896 ignition[829]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 May 15 15:12:49.866715 ignition[829]: GET result: OK May 15 15:12:49.866938 ignition[829]: parsing config with SHA512: 24b5fbb72c336ad1b5bcc8d205e24ef78a62effcf96c77a7311b4c64e19b1d9442306a744bec4f3165446487eae602420c37654827ccd95e330a6c6677bd5e71 May 15 15:12:49.872454 unknown[829]: fetched base config from "system" May 15 15:12:49.872466 unknown[829]: fetched base config from "system" May 15 15:12:49.872472 unknown[829]: fetched user config from "digitalocean" May 15 15:12:49.873402 ignition[829]: fetch: fetch complete May 15 15:12:49.873407 ignition[829]: fetch: fetch passed May 15 15:12:49.873460 ignition[829]: Ignition finished successfully May 15 15:12:49.876905 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). May 15 15:12:49.878517 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 15 15:12:49.918358 ignition[836]: Ignition 2.21.0 May 15 15:12:49.918370 ignition[836]: Stage: kargs May 15 15:12:49.918568 ignition[836]: no configs at "/usr/lib/ignition/base.d" May 15 15:12:49.918579 ignition[836]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 15:12:49.921209 ignition[836]: kargs: kargs passed May 15 15:12:49.921330 ignition[836]: Ignition finished successfully May 15 15:12:49.923953 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 15 15:12:49.925776 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 15 15:12:49.962834 ignition[843]: Ignition 2.21.0 May 15 15:12:49.962851 ignition[843]: Stage: disks May 15 15:12:49.963164 ignition[843]: no configs at "/usr/lib/ignition/base.d" May 15 15:12:49.963176 ignition[843]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 15:12:49.964527 ignition[843]: disks: disks passed May 15 15:12:49.964587 ignition[843]: Ignition finished successfully May 15 15:12:49.966407 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 15 15:12:49.967098 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 15 15:12:49.967540 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 15 15:12:49.968374 systemd[1]: Reached target local-fs.target - Local File Systems. May 15 15:12:49.969094 systemd[1]: Reached target sysinit.target - System Initialization. May 15 15:12:49.969641 systemd[1]: Reached target basic.target - Basic System. May 15 15:12:49.971483 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 15 15:12:50.002487 systemd-fsck[852]: ROOT: clean, 15/553520 files, 52789/553472 blocks May 15 15:12:50.005306 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 15 15:12:50.007667 systemd[1]: Mounting sysroot.mount - /sysroot... May 15 15:12:50.119981 kernel: EXT4-fs (vda9): mounted filesystem f7dea4bd-2644-4592-b85b-330f322c4d2b r/w with ordered data mode. Quota mode: none. May 15 15:12:50.120762 systemd[1]: Mounted sysroot.mount - /sysroot. May 15 15:12:50.122157 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 15 15:12:50.124803 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 15 15:12:50.128067 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 15 15:12:50.129708 systemd[1]: Starting flatcar-afterburn-network.service - Flatcar Afterburn network service... May 15 15:12:50.136530 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... May 15 15:12:50.139110 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 15 15:12:50.140096 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 15 15:12:50.143682 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 15 15:12:50.147109 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 15 15:12:50.161196 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (860) May 15 15:12:50.165163 kernel: BTRFS info (device vda6): first mount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 15:12:50.165225 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 15 15:12:50.165239 kernel: BTRFS info (device vda6): using free-space-tree May 15 15:12:50.178445 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 15 15:12:50.230988 coreos-metadata[863]: May 15 15:12:50.230 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 15 15:12:50.233108 coreos-metadata[862]: May 15 15:12:50.232 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 15 15:12:50.235830 initrd-setup-root[890]: cut: /sysroot/etc/passwd: No such file or directory May 15 15:12:50.242253 initrd-setup-root[897]: cut: /sysroot/etc/group: No such file or directory May 15 15:12:50.244771 coreos-metadata[863]: May 15 15:12:50.243 INFO Fetch successful May 15 15:12:50.245316 coreos-metadata[862]: May 15 15:12:50.245 INFO Fetch successful May 15 15:12:50.251939 systemd[1]: flatcar-afterburn-network.service: Deactivated successfully. May 15 15:12:50.252759 systemd[1]: Finished flatcar-afterburn-network.service - Flatcar Afterburn network service. May 15 15:12:50.254381 coreos-metadata[863]: May 15 15:12:50.254 INFO wrote hostname ci-4334.0.0-a-073c24674a to /sysroot/etc/hostname May 15 15:12:50.255582 initrd-setup-root[905]: cut: /sysroot/etc/shadow: No such file or directory May 15 15:12:50.256649 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 15 15:12:50.263175 initrd-setup-root[913]: cut: /sysroot/etc/gshadow: No such file or directory May 15 15:12:50.367004 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 15 15:12:50.369565 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 15 15:12:50.371173 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 15 15:12:50.385990 kernel: BTRFS info (device vda6): last unmount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 15:12:50.404774 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 15 15:12:50.417641 ignition[984]: INFO : Ignition 2.21.0 May 15 15:12:50.417641 ignition[984]: INFO : Stage: mount May 15 15:12:50.419909 ignition[984]: INFO : no configs at "/usr/lib/ignition/base.d" May 15 15:12:50.419909 ignition[984]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 15:12:50.419909 ignition[984]: INFO : mount: mount passed May 15 15:12:50.419909 ignition[984]: INFO : Ignition finished successfully May 15 15:12:50.421186 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 15 15:12:50.423185 systemd[1]: Starting ignition-files.service - Ignition (files)... May 15 15:12:50.522420 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 15 15:12:50.524142 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 15 15:12:50.549081 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (996) May 15 15:12:50.549143 kernel: BTRFS info (device vda6): first mount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 15:12:50.549158 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 15 15:12:50.550137 kernel: BTRFS info (device vda6): using free-space-tree May 15 15:12:50.556608 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 15 15:12:50.588982 ignition[1013]: INFO : Ignition 2.21.0 May 15 15:12:50.588982 ignition[1013]: INFO : Stage: files May 15 15:12:50.588982 ignition[1013]: INFO : no configs at "/usr/lib/ignition/base.d" May 15 15:12:50.588982 ignition[1013]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 15:12:50.594760 ignition[1013]: DEBUG : files: compiled without relabeling support, skipping May 15 15:12:50.596612 ignition[1013]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 15 15:12:50.596612 ignition[1013]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 15 15:12:50.599656 ignition[1013]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 15 15:12:50.600170 ignition[1013]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 15 15:12:50.600623 ignition[1013]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 15 15:12:50.600177 unknown[1013]: wrote ssh authorized keys file for user: core May 15 15:12:50.602043 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" May 15 15:12:50.602657 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 May 15 15:12:50.655446 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 15 15:12:50.940257 systemd-networkd[820]: eth1: Gained IPv6LL May 15 15:12:50.979794 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" May 15 15:12:50.980532 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" May 15 15:12:50.980532 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" May 15 15:12:50.980532 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" May 15 15:12:50.980532 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" May 15 15:12:50.980532 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 15 15:12:50.980532 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 15 15:12:50.980532 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 15 15:12:50.980532 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 15 15:12:50.989474 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" May 15 15:12:50.989474 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 15 15:12:50.989474 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 15 15:12:50.989474 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 15 15:12:50.989474 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 15 15:12:50.989474 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 May 15 15:12:51.260177 systemd-networkd[820]: eth0: Gained IPv6LL May 15 15:12:51.480436 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK May 15 15:12:51.712765 ignition[1013]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 15 15:12:51.713938 ignition[1013]: INFO : files: op(b): [started] processing unit "prepare-helm.service" May 15 15:12:51.715283 ignition[1013]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 15 15:12:51.716996 ignition[1013]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 15 15:12:51.718232 ignition[1013]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" May 15 15:12:51.718232 ignition[1013]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" May 15 15:12:51.718232 ignition[1013]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" May 15 15:12:51.718232 ignition[1013]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" May 15 15:12:51.718232 ignition[1013]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" May 15 15:12:51.718232 ignition[1013]: INFO : files: files passed May 15 15:12:51.718232 ignition[1013]: INFO : Ignition finished successfully May 15 15:12:51.721267 systemd[1]: Finished ignition-files.service - Ignition (files). May 15 15:12:51.722636 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 15 15:12:51.725207 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 15 15:12:51.746292 systemd[1]: ignition-quench.service: Deactivated successfully. May 15 15:12:51.747324 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 15 15:12:51.754438 initrd-setup-root-after-ignition[1042]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 15 15:12:51.754438 initrd-setup-root-after-ignition[1042]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 15 15:12:51.755837 initrd-setup-root-after-ignition[1046]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 15 15:12:51.757725 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 15 15:12:51.758476 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 15 15:12:51.760017 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 15 15:12:51.817582 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 15 15:12:51.817709 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 15 15:12:51.818704 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 15 15:12:51.819194 systemd[1]: Reached target initrd.target - Initrd Default Target. May 15 15:12:51.819857 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 15 15:12:51.820827 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 15 15:12:51.847679 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 15 15:12:51.850461 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 15 15:12:51.874492 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 15 15:12:51.875538 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 15 15:12:51.875974 systemd[1]: Stopped target timers.target - Timer Units. May 15 15:12:51.876376 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 15 15:12:51.876513 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 15 15:12:51.877591 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 15 15:12:51.878335 systemd[1]: Stopped target basic.target - Basic System. May 15 15:12:51.879312 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 15 15:12:51.880189 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 15 15:12:51.880825 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 15 15:12:51.881708 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. May 15 15:12:51.882736 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 15 15:12:51.883925 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 15 15:12:51.886561 systemd[1]: Stopped target sysinit.target - System Initialization. May 15 15:12:51.887438 systemd[1]: Stopped target local-fs.target - Local File Systems. May 15 15:12:51.888321 systemd[1]: Stopped target swap.target - Swaps. May 15 15:12:51.889115 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 15 15:12:51.889349 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 15 15:12:51.890392 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 15 15:12:51.891111 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 15 15:12:51.891812 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 15 15:12:51.891955 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 15 15:12:51.892492 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 15 15:12:51.892644 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 15 15:12:51.893531 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 15 15:12:51.893733 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 15 15:12:51.894429 systemd[1]: ignition-files.service: Deactivated successfully. May 15 15:12:51.894569 systemd[1]: Stopped ignition-files.service - Ignition (files). May 15 15:12:51.895496 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. May 15 15:12:51.895692 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 15 15:12:51.899083 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 15 15:12:51.901179 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 15 15:12:51.901530 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 15 15:12:51.901647 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 15 15:12:51.903421 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 15 15:12:51.903574 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 15 15:12:51.913764 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 15 15:12:51.915097 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 15 15:12:51.935554 ignition[1066]: INFO : Ignition 2.21.0 May 15 15:12:51.935554 ignition[1066]: INFO : Stage: umount May 15 15:12:51.937218 ignition[1066]: INFO : no configs at "/usr/lib/ignition/base.d" May 15 15:12:51.937218 ignition[1066]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 15:12:51.937218 ignition[1066]: INFO : umount: umount passed May 15 15:12:51.937218 ignition[1066]: INFO : Ignition finished successfully May 15 15:12:51.938806 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 15 15:12:51.939489 systemd[1]: ignition-mount.service: Deactivated successfully. May 15 15:12:51.945072 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 15 15:12:51.990266 systemd[1]: ignition-disks.service: Deactivated successfully. May 15 15:12:51.990413 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 15 15:12:51.991307 systemd[1]: ignition-kargs.service: Deactivated successfully. May 15 15:12:51.991361 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 15 15:12:51.993467 systemd[1]: ignition-fetch.service: Deactivated successfully. May 15 15:12:51.993531 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). May 15 15:12:51.993884 systemd[1]: Stopped target network.target - Network. May 15 15:12:51.995088 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 15 15:12:51.995153 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 15 15:12:51.995909 systemd[1]: Stopped target paths.target - Path Units. May 15 15:12:51.996494 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 15 15:12:52.000060 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 15 15:12:52.000582 systemd[1]: Stopped target slices.target - Slice Units. May 15 15:12:52.001400 systemd[1]: Stopped target sockets.target - Socket Units. May 15 15:12:52.002142 systemd[1]: iscsid.socket: Deactivated successfully. May 15 15:12:52.002209 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 15 15:12:52.002747 systemd[1]: iscsiuio.socket: Deactivated successfully. May 15 15:12:52.002797 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 15 15:12:52.003364 systemd[1]: ignition-setup.service: Deactivated successfully. May 15 15:12:52.003443 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 15 15:12:52.004033 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 15 15:12:52.004086 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 15 15:12:52.004879 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 15 15:12:52.005572 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 15 15:12:52.006928 systemd[1]: sysroot-boot.service: Deactivated successfully. May 15 15:12:52.007094 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 15 15:12:52.008292 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 15 15:12:52.008453 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 15 15:12:52.011736 systemd[1]: systemd-resolved.service: Deactivated successfully. May 15 15:12:52.011840 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 15 15:12:52.015056 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 15 15:12:52.016320 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 15 15:12:52.016393 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 15 15:12:52.018516 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 15 15:12:52.019855 systemd[1]: systemd-networkd.service: Deactivated successfully. May 15 15:12:52.020556 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 15 15:12:52.022684 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 15 15:12:52.023545 systemd[1]: Stopped target network-pre.target - Preparation for Network. May 15 15:12:52.024469 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 15 15:12:52.024515 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 15 15:12:52.025556 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 15 15:12:52.027621 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 15 15:12:52.028095 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 15 15:12:52.028544 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 15 15:12:52.028596 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 15 15:12:52.030487 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 15 15:12:52.030538 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 15 15:12:52.031812 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 15 15:12:52.034672 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 15 15:12:52.047470 systemd[1]: systemd-udevd.service: Deactivated successfully. May 15 15:12:52.049153 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 15 15:12:52.050311 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 15 15:12:52.050391 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 15 15:12:52.051204 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 15 15:12:52.051239 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 15 15:12:52.051931 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 15 15:12:52.051997 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 15 15:12:52.053123 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 15 15:12:52.053168 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 15 15:12:52.054295 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 15 15:12:52.054349 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 15 15:12:52.056131 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 15 15:12:52.057191 systemd[1]: systemd-network-generator.service: Deactivated successfully. May 15 15:12:52.057257 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. May 15 15:12:52.058765 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 15 15:12:52.058812 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 15 15:12:52.060936 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. May 15 15:12:52.061001 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 15 15:12:52.063062 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 15 15:12:52.063107 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 15 15:12:52.063855 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 15 15:12:52.063914 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 15 15:12:52.066341 systemd[1]: network-cleanup.service: Deactivated successfully. May 15 15:12:52.069272 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 15 15:12:52.077180 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 15 15:12:52.077331 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 15 15:12:52.078698 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 15 15:12:52.080439 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 15 15:12:52.103252 systemd[1]: Switching root. May 15 15:12:52.141384 systemd-journald[211]: Journal stopped May 15 15:12:53.353181 systemd-journald[211]: Received SIGTERM from PID 1 (systemd). May 15 15:12:53.353268 kernel: SELinux: policy capability network_peer_controls=1 May 15 15:12:53.353289 kernel: SELinux: policy capability open_perms=1 May 15 15:12:53.353301 kernel: SELinux: policy capability extended_socket_class=1 May 15 15:12:53.353318 kernel: SELinux: policy capability always_check_network=0 May 15 15:12:53.353336 kernel: SELinux: policy capability cgroup_seclabel=1 May 15 15:12:53.353348 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 15 15:12:53.353361 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 15 15:12:53.353372 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 15 15:12:53.353383 kernel: SELinux: policy capability userspace_initial_context=0 May 15 15:12:53.353395 kernel: audit: type=1403 audit(1747321972.288:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 15 15:12:53.353407 systemd[1]: Successfully loaded SELinux policy in 47.509ms. May 15 15:12:53.353434 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 13.932ms. May 15 15:12:53.353451 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 15 15:12:53.353464 systemd[1]: Detected virtualization kvm. May 15 15:12:53.353476 systemd[1]: Detected architecture x86-64. May 15 15:12:53.353488 systemd[1]: Detected first boot. May 15 15:12:53.353501 systemd[1]: Hostname set to . May 15 15:12:53.353513 systemd[1]: Initializing machine ID from VM UUID. May 15 15:12:53.353525 zram_generator::config[1110]: No configuration found. May 15 15:12:53.353541 kernel: Guest personality initialized and is inactive May 15 15:12:53.353553 kernel: VMCI host device registered (name=vmci, major=10, minor=125) May 15 15:12:53.353569 kernel: Initialized host personality May 15 15:12:53.353580 kernel: NET: Registered PF_VSOCK protocol family May 15 15:12:53.353591 systemd[1]: Populated /etc with preset unit settings. May 15 15:12:53.353605 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 15 15:12:53.353618 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 15 15:12:53.353634 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 15 15:12:53.353646 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 15 15:12:53.353662 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 15 15:12:53.353677 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 15 15:12:53.353717 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 15 15:12:53.353730 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 15 15:12:53.353743 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 15 15:12:53.353757 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 15 15:12:53.353769 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 15 15:12:53.353781 systemd[1]: Created slice user.slice - User and Session Slice. May 15 15:12:53.353799 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 15 15:12:53.353811 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 15 15:12:53.353823 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 15 15:12:53.353836 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 15 15:12:53.353849 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 15 15:12:53.353861 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 15 15:12:53.353874 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 15 15:12:53.353888 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 15 15:12:53.353901 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 15 15:12:53.353913 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 15 15:12:53.353925 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 15 15:12:53.353937 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 15 15:12:53.353949 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 15 15:12:53.366464 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 15 15:12:53.366505 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 15 15:12:53.366520 systemd[1]: Reached target slices.target - Slice Units. May 15 15:12:53.366543 systemd[1]: Reached target swap.target - Swaps. May 15 15:12:53.366555 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 15 15:12:53.366568 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 15 15:12:53.366581 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 15 15:12:53.366598 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 15 15:12:53.366611 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 15 15:12:53.366623 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 15 15:12:53.366635 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 15 15:12:53.366647 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 15 15:12:53.366662 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 15 15:12:53.366674 systemd[1]: Mounting media.mount - External Media Directory... May 15 15:12:53.366687 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:53.366699 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 15 15:12:53.366712 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 15 15:12:53.366724 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 15 15:12:53.366737 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 15 15:12:53.366754 systemd[1]: Reached target machines.target - Containers. May 15 15:12:53.366769 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 15 15:12:53.366782 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 15:12:53.366794 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 15 15:12:53.366806 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 15 15:12:53.366819 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 15 15:12:53.366831 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 15 15:12:53.366843 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 15 15:12:53.366855 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 15 15:12:53.366867 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 15 15:12:53.366890 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 15 15:12:53.366908 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 15 15:12:53.366927 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 15 15:12:53.366940 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 15 15:12:53.366952 systemd[1]: Stopped systemd-fsck-usr.service. May 15 15:12:53.366984 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 15:12:53.366998 systemd[1]: Starting systemd-journald.service - Journal Service... May 15 15:12:53.367010 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 15 15:12:53.367028 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 15 15:12:53.367041 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 15 15:12:53.367054 kernel: loop: module loaded May 15 15:12:53.367069 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 15 15:12:53.367081 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 15 15:12:53.367096 systemd[1]: verity-setup.service: Deactivated successfully. May 15 15:12:53.367109 systemd[1]: Stopped verity-setup.service. May 15 15:12:53.367130 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:53.367143 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 15 15:12:53.367156 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 15 15:12:53.367168 systemd[1]: Mounted media.mount - External Media Directory. May 15 15:12:53.367184 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 15 15:12:53.367197 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 15 15:12:53.367210 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 15 15:12:53.367222 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 15 15:12:53.367235 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 15 15:12:53.367249 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 15 15:12:53.367262 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 15 15:12:53.367275 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 15 15:12:53.367294 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 15 15:12:53.367307 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 15 15:12:53.367320 systemd[1]: modprobe@loop.service: Deactivated successfully. May 15 15:12:53.367332 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 15 15:12:53.367345 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 15 15:12:53.367360 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 15 15:12:53.367373 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 15 15:12:53.367385 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 15 15:12:53.367398 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 15 15:12:53.367414 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 15 15:12:53.367427 systemd[1]: Reached target local-fs.target - Local File Systems. May 15 15:12:53.367439 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 15 15:12:53.367453 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 15 15:12:53.367468 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 15:12:53.367480 kernel: fuse: init (API version 7.41) May 15 15:12:53.367492 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 15 15:12:53.367506 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 15 15:12:53.367519 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 15 15:12:53.367534 kernel: ACPI: bus type drm_connector registered May 15 15:12:53.367546 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 15 15:12:53.367558 systemd[1]: modprobe@drm.service: Deactivated successfully. May 15 15:12:53.367571 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 15 15:12:53.367589 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 15 15:12:53.367601 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 15 15:12:53.367614 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 15 15:12:53.367628 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 15 15:12:53.367641 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 15 15:12:53.367657 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 15 15:12:53.367670 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 15 15:12:53.367682 systemd[1]: Reached target network-pre.target - Preparation for Network. May 15 15:12:53.367695 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 15 15:12:53.367708 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 15 15:12:53.367721 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 15 15:12:53.367781 systemd-journald[1187]: Collecting audit messages is disabled. May 15 15:12:53.367812 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 15 15:12:53.367828 kernel: loop0: detected capacity change from 0 to 8 May 15 15:12:53.367841 systemd-journald[1187]: Journal started May 15 15:12:53.367866 systemd-journald[1187]: Runtime Journal (/run/log/journal/782f0142214d40feb3e3035039648a79) is 4.9M, max 39.5M, 34.6M free. May 15 15:12:53.373561 systemd[1]: Started systemd-journald.service - Journal Service. May 15 15:12:52.902869 systemd[1]: Queued start job for default target multi-user.target. May 15 15:12:52.910953 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. May 15 15:12:52.911595 systemd[1]: systemd-journald.service: Deactivated successfully. May 15 15:12:53.357163 systemd-tmpfiles[1198]: ACLs are not supported, ignoring. May 15 15:12:53.357187 systemd-tmpfiles[1198]: ACLs are not supported, ignoring. May 15 15:12:53.411715 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 15 15:12:53.379247 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 15 15:12:53.390612 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 15 15:12:53.407187 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 15 15:12:53.412048 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 15 15:12:53.419638 kernel: loop1: detected capacity change from 0 to 146240 May 15 15:12:53.419174 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 15 15:12:53.433414 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 15 15:12:53.443098 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 15 15:12:53.461278 systemd-journald[1187]: Time spent on flushing to /var/log/journal/782f0142214d40feb3e3035039648a79 is 58.953ms for 1021 entries. May 15 15:12:53.461278 systemd-journald[1187]: System Journal (/var/log/journal/782f0142214d40feb3e3035039648a79) is 8M, max 195.6M, 187.6M free. May 15 15:12:53.525285 systemd-journald[1187]: Received client request to flush runtime journal. May 15 15:12:53.525329 kernel: loop2: detected capacity change from 0 to 218376 May 15 15:12:53.525344 kernel: loop3: detected capacity change from 0 to 113872 May 15 15:12:53.527411 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 15 15:12:53.557274 kernel: loop4: detected capacity change from 0 to 8 May 15 15:12:53.561483 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 15 15:12:53.566172 kernel: loop5: detected capacity change from 0 to 146240 May 15 15:12:53.569218 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 15 15:12:53.596725 kernel: loop6: detected capacity change from 0 to 218376 May 15 15:12:53.621985 kernel: loop7: detected capacity change from 0 to 113872 May 15 15:12:53.643469 (sd-merge)[1258]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. May 15 15:12:53.645265 (sd-merge)[1258]: Merged extensions into '/usr'. May 15 15:12:53.648030 systemd-tmpfiles[1260]: ACLs are not supported, ignoring. May 15 15:12:53.648048 systemd-tmpfiles[1260]: ACLs are not supported, ignoring. May 15 15:12:53.655076 systemd[1]: Reload requested from client PID 1209 ('systemd-sysext') (unit systemd-sysext.service)... May 15 15:12:53.655092 systemd[1]: Reloading... May 15 15:12:53.814032 zram_generator::config[1288]: No configuration found. May 15 15:12:53.952857 ldconfig[1204]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 15 15:12:53.962572 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 15 15:12:54.054543 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 15 15:12:54.054751 systemd[1]: Reloading finished in 399 ms. May 15 15:12:54.073642 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 15 15:12:54.074446 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 15 15:12:54.075215 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 15 15:12:54.085295 systemd[1]: Starting ensure-sysext.service... May 15 15:12:54.091226 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 15 15:12:54.126098 systemd[1]: Reload requested from client PID 1332 ('systemctl') (unit ensure-sysext.service)... May 15 15:12:54.126115 systemd[1]: Reloading... May 15 15:12:54.164203 systemd-tmpfiles[1333]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. May 15 15:12:54.165802 systemd-tmpfiles[1333]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. May 15 15:12:54.166182 systemd-tmpfiles[1333]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 15 15:12:54.166430 systemd-tmpfiles[1333]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 15 15:12:54.169601 systemd-tmpfiles[1333]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 15 15:12:54.171212 systemd-tmpfiles[1333]: ACLs are not supported, ignoring. May 15 15:12:54.171360 systemd-tmpfiles[1333]: ACLs are not supported, ignoring. May 15 15:12:54.179362 systemd-tmpfiles[1333]: Detected autofs mount point /boot during canonicalization of boot. May 15 15:12:54.181116 systemd-tmpfiles[1333]: Skipping /boot May 15 15:12:54.214289 systemd-tmpfiles[1333]: Detected autofs mount point /boot during canonicalization of boot. May 15 15:12:54.214475 systemd-tmpfiles[1333]: Skipping /boot May 15 15:12:54.228091 zram_generator::config[1356]: No configuration found. May 15 15:12:54.401092 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 15 15:12:54.499892 systemd[1]: Reloading finished in 373 ms. May 15 15:12:54.523312 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 15 15:12:54.529576 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 15 15:12:54.540160 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 15 15:12:54.543639 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 15 15:12:54.548095 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 15 15:12:54.557057 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 15 15:12:54.559136 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 15 15:12:54.564433 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 15 15:12:54.570675 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:54.570861 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 15:12:54.576456 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 15 15:12:54.585201 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 15 15:12:54.588341 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 15 15:12:54.588793 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 15:12:54.588909 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 15:12:54.589020 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:54.592787 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:54.592990 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 15:12:54.593207 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 15:12:54.593334 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 15:12:54.593481 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:54.599291 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:54.599543 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 15:12:54.607318 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 15 15:12:54.607823 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 15:12:54.607936 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 15:12:54.608089 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:54.620121 systemd[1]: Finished ensure-sysext.service. May 15 15:12:54.635120 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... May 15 15:12:54.642378 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 15 15:12:54.645186 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 15 15:12:54.646020 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 15 15:12:54.646810 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 15 15:12:54.647735 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 15 15:12:54.648776 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 15 15:12:54.649680 systemd[1]: modprobe@loop.service: Deactivated successfully. May 15 15:12:54.651237 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 15 15:12:54.656386 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 15 15:12:54.657909 systemd[1]: modprobe@drm.service: Deactivated successfully. May 15 15:12:54.659117 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 15 15:12:54.667751 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 15 15:12:54.667856 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 15 15:12:54.667893 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 15 15:12:54.681647 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 15 15:12:54.684505 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 15 15:12:54.704765 systemd-udevd[1409]: Using default interface naming scheme 'v255'. May 15 15:12:54.715152 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 15 15:12:54.729462 augenrules[1446]: No rules May 15 15:12:54.733356 systemd[1]: audit-rules.service: Deactivated successfully. May 15 15:12:54.734222 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 15 15:12:54.734776 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 15 15:12:54.741007 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 15 15:12:54.754156 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 15 15:12:54.894435 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. May 15 15:12:54.928034 systemd[1]: Condition check resulted in dev-disk-by\x2dlabel-config\x2d2.device - /dev/disk/by-label/config-2 being skipped. May 15 15:12:54.930188 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... May 15 15:12:54.930553 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:54.930693 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 15:12:54.934153 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 15 15:12:54.941472 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 15 15:12:54.947430 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 15 15:12:54.947892 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 15:12:54.947937 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 15:12:54.947981 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 15 15:12:54.947998 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 15:12:54.948475 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 15 15:12:54.948716 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 15 15:12:54.967305 systemd-resolved[1408]: Positive Trust Anchors: May 15 15:12:54.967318 systemd-resolved[1408]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 15 15:12:54.967355 systemd-resolved[1408]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 15 15:12:54.973387 systemd-resolved[1408]: Using system hostname 'ci-4334.0.0-a-073c24674a'. May 15 15:12:54.975166 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 15 15:12:54.975705 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 15 15:12:54.999083 kernel: ISO 9660 Extensions: RRIP_1991A May 15 15:12:55.037341 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. May 15 15:12:55.040331 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 15 15:12:55.040522 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 15 15:12:55.041227 systemd[1]: modprobe@loop.service: Deactivated successfully. May 15 15:12:55.041393 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 15 15:12:55.045899 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 15 15:12:55.046646 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 15 15:12:55.053397 systemd-networkd[1455]: lo: Link UP May 15 15:12:55.053410 systemd-networkd[1455]: lo: Gained carrier May 15 15:12:55.055935 systemd-networkd[1455]: Enumeration completed May 15 15:12:55.056440 systemd-networkd[1455]: eth0: Configuring with /run/systemd/network/10-9e:7e:9c:b4:68:fe.network. May 15 15:12:55.057183 systemd-networkd[1455]: eth0: Link UP May 15 15:12:55.057364 systemd[1]: Started systemd-networkd.service - Network Configuration. May 15 15:12:55.057515 systemd-networkd[1455]: eth0: Gained carrier May 15 15:12:55.058637 systemd[1]: Reached target network.target - Network. May 15 15:12:55.061076 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 15 15:12:55.064195 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 15 15:12:55.080105 systemd-networkd[1455]: eth1: Configuring with /run/systemd/network/10-8e:d7:ca:74:6e:ce.network. May 15 15:12:55.080656 systemd-networkd[1455]: eth1: Link UP May 15 15:12:55.082227 systemd-networkd[1455]: eth1: Gained carrier May 15 15:12:55.114347 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. May 15 15:12:55.114829 systemd[1]: Reached target sysinit.target - System Initialization. May 15 15:12:55.116038 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 15 15:12:55.116447 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 15 15:12:55.117140 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. May 15 15:12:55.118033 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 15 15:12:55.119055 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 15 15:12:55.119093 systemd[1]: Reached target paths.target - Path Units. May 15 15:12:55.119412 systemd[1]: Reached target time-set.target - System Time Set. May 15 15:12:55.119864 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 15 15:12:55.121027 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 15 15:12:55.121401 systemd[1]: Reached target timers.target - Timer Units. May 15 15:12:55.125327 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 15 15:12:55.127894 systemd[1]: Starting docker.socket - Docker Socket for the API... May 15 15:12:55.137766 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 15 15:12:55.139387 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 15 15:12:55.140915 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 15 15:12:55.151067 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 15 15:12:55.152749 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 15 15:12:55.157045 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 15 15:12:55.157728 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 15 15:12:55.178469 systemd[1]: Reached target sockets.target - Socket Units. May 15 15:12:55.179053 kernel: mousedev: PS/2 mouse device common for all mice May 15 15:12:55.179288 systemd[1]: Reached target basic.target - Basic System. May 15 15:12:55.180097 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 15 15:12:55.180131 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 15 15:12:55.182277 systemd[1]: Starting containerd.service - containerd container runtime... May 15 15:12:55.185476 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... May 15 15:12:55.191985 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 May 15 15:12:55.188420 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 15 15:12:55.193230 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 15 15:12:55.196236 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 15 15:12:55.199239 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 15 15:12:55.200076 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 15 15:12:55.208302 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... May 15 15:12:55.214743 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 15 15:12:55.223903 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 15 15:12:55.229902 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 15 15:12:55.235284 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 15 15:12:55.248007 systemd[1]: Starting systemd-logind.service - User Login Management... May 15 15:12:55.251302 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 15 15:12:55.257642 kernel: ACPI: button: Power Button [PWRF] May 15 15:12:55.257141 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 15 15:12:55.260776 google_oslogin_nss_cache[1519]: oslogin_cache_refresh[1519]: Refreshing passwd entry cache May 15 15:12:55.260389 systemd[1]: Starting update-engine.service - Update Engine... May 15 15:12:55.259723 oslogin_cache_refresh[1519]: Refreshing passwd entry cache May 15 15:12:55.269984 jq[1517]: false May 15 15:12:55.265305 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 15 15:12:55.276292 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 15 15:12:55.277309 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 15 15:12:55.278236 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 15 15:12:55.278542 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 15 15:12:55.278707 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 15 15:12:55.294082 google_oslogin_nss_cache[1519]: oslogin_cache_refresh[1519]: Failure getting users, quitting May 15 15:12:55.294082 google_oslogin_nss_cache[1519]: oslogin_cache_refresh[1519]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 15 15:12:55.294082 google_oslogin_nss_cache[1519]: oslogin_cache_refresh[1519]: Refreshing group entry cache May 15 15:12:55.294082 google_oslogin_nss_cache[1519]: oslogin_cache_refresh[1519]: Failure getting groups, quitting May 15 15:12:55.294082 google_oslogin_nss_cache[1519]: oslogin_cache_refresh[1519]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 15 15:12:55.287247 oslogin_cache_refresh[1519]: Failure getting users, quitting May 15 15:12:55.287267 oslogin_cache_refresh[1519]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 15 15:12:55.287323 oslogin_cache_refresh[1519]: Refreshing group entry cache May 15 15:12:55.292687 oslogin_cache_refresh[1519]: Failure getting groups, quitting May 15 15:12:55.292703 oslogin_cache_refresh[1519]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 15 15:12:55.296101 systemd[1]: google-oslogin-cache.service: Deactivated successfully. May 15 15:12:55.297113 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. May 15 15:12:55.310943 jq[1529]: true May 15 15:12:55.339157 extend-filesystems[1518]: Found loop4 May 15 15:12:55.339157 extend-filesystems[1518]: Found loop5 May 15 15:12:55.339157 extend-filesystems[1518]: Found loop6 May 15 15:12:55.339157 extend-filesystems[1518]: Found loop7 May 15 15:12:55.339157 extend-filesystems[1518]: Found vda May 15 15:12:55.339157 extend-filesystems[1518]: Found vda1 May 15 15:12:55.339157 extend-filesystems[1518]: Found vda2 May 15 15:12:55.339157 extend-filesystems[1518]: Found vda3 May 15 15:12:55.339157 extend-filesystems[1518]: Found usr May 15 15:12:55.339157 extend-filesystems[1518]: Found vda4 May 15 15:12:55.339157 extend-filesystems[1518]: Found vda6 May 15 15:12:55.339157 extend-filesystems[1518]: Found vda7 May 15 15:12:55.339157 extend-filesystems[1518]: Found vda9 May 15 15:12:55.339157 extend-filesystems[1518]: Checking size of /dev/vda9 May 15 15:12:55.390398 jq[1541]: true May 15 15:12:55.390705 coreos-metadata[1513]: May 15 15:12:55.343 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 15 15:12:55.390705 coreos-metadata[1513]: May 15 15:12:55.358 INFO Fetch successful May 15 15:12:55.396710 tar[1538]: linux-amd64/LICENSE May 15 15:12:55.396710 tar[1538]: linux-amd64/helm May 15 15:12:55.400887 update_engine[1528]: I20250515 15:12:55.400779 1528 main.cc:92] Flatcar Update Engine starting May 15 15:12:55.412265 extend-filesystems[1518]: Resized partition /dev/vda9 May 15 15:12:55.415245 (ntainerd)[1549]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 15 15:12:55.433109 dbus-daemon[1514]: [system] SELinux support is enabled May 15 15:12:55.433333 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 15 15:12:55.437819 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 15 15:12:55.441766 extend-filesystems[1568]: resize2fs 1.47.2 (1-Jan-2025) May 15 15:12:55.437869 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 15 15:12:55.439218 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 15 15:12:55.439296 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). May 15 15:12:55.439315 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 15 15:12:55.449323 systemd[1]: Started update-engine.service - Update Engine. May 15 15:12:55.459899 update_engine[1528]: I20250515 15:12:55.454250 1528 update_check_scheduler.cc:74] Next update check in 11m36s May 15 15:12:55.460052 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks May 15 15:12:55.863536 systemd-timesyncd[1424]: Contacted time server 23.186.168.123:123 (0.flatcar.pool.ntp.org). May 15 15:12:55.863608 systemd-timesyncd[1424]: Initial clock synchronization to Thu 2025-05-15 15:12:55.863304 UTC. May 15 15:12:55.865448 systemd-resolved[1408]: Clock change detected. Flushing caches. May 15 15:12:55.877839 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 15 15:12:55.881581 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 15 15:12:55.888674 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 15 15:12:55.896258 systemd[1]: motdgen.service: Deactivated successfully. May 15 15:12:55.896500 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 15 15:12:55.934616 bash[1575]: Updated "/home/core/.ssh/authorized_keys" May 15 15:12:55.933944 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 15 15:12:55.946337 systemd[1]: Starting sshkeys.service... May 15 15:12:55.947756 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. May 15 15:12:55.951622 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 15 15:12:56.016823 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. May 15 15:12:56.022043 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... May 15 15:12:56.035885 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 15 15:12:56.068547 systemd-logind[1527]: New seat seat0. May 15 15:12:56.094654 systemd[1]: Started systemd-logind.service - User Login Management. May 15 15:12:56.101042 kernel: EXT4-fs (vda9): resized filesystem to 15121403 May 15 15:12:56.124808 extend-filesystems[1568]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 15 15:12:56.124808 extend-filesystems[1568]: old_desc_blocks = 1, new_desc_blocks = 8 May 15 15:12:56.124808 extend-filesystems[1568]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. May 15 15:12:56.133787 extend-filesystems[1518]: Resized filesystem in /dev/vda9 May 15 15:12:56.133787 extend-filesystems[1518]: Found vdb May 15 15:12:56.126240 systemd[1]: extend-filesystems.service: Deactivated successfully. May 15 15:12:56.127431 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 15 15:12:56.147464 coreos-metadata[1586]: May 15 15:12:56.147 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 15 15:12:56.166595 coreos-metadata[1586]: May 15 15:12:56.160 INFO Fetch successful May 15 15:12:56.187502 unknown[1586]: wrote ssh authorized keys file for user: core May 15 15:12:56.216369 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). May 15 15:12:56.217975 update-ssh-keys[1607]: Updated "/home/core/.ssh/authorized_keys" May 15 15:12:56.222430 systemd[1]: Finished sshkeys.service. May 15 15:12:56.247952 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 May 15 15:12:56.310949 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD May 15 15:12:56.271102 locksmithd[1574]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 15 15:12:56.446310 sshd_keygen[1543]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 15 15:12:56.457992 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 May 15 15:12:56.458075 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console May 15 15:12:56.462954 kernel: Console: switching to colour dummy device 80x25 May 15 15:12:56.464952 kernel: [drm] features: -virgl +edid -resource_blob -host_visible May 15 15:12:56.465067 kernel: [drm] features: -context_init May 15 15:12:56.465952 kernel: [drm] number of scanouts: 1 May 15 15:12:56.466955 kernel: [drm] number of cap sets: 0 May 15 15:12:56.468949 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:02.0 on minor 0 May 15 15:12:56.533514 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 15 15:12:56.535720 systemd[1]: Starting issuegen.service - Generate /run/issue... May 15 15:12:56.564084 systemd[1]: issuegen.service: Deactivated successfully. May 15 15:12:56.564830 systemd[1]: Finished issuegen.service - Generate /run/issue. May 15 15:12:56.568020 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 15 15:12:56.588276 systemd-logind[1527]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 15 15:12:56.606373 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 15 15:12:56.617199 containerd[1549]: time="2025-05-15T15:12:56Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 15 15:12:56.625862 containerd[1549]: time="2025-05-15T15:12:56.625796981Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 May 15 15:12:56.634724 systemd-logind[1527]: Watching system buttons on /dev/input/event2 (Power Button) May 15 15:12:56.699829 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 15 15:12:56.707555 systemd[1]: Started getty@tty1.service - Getty on tty1. May 15 15:12:56.712039 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 15 15:12:56.712256 systemd[1]: Reached target getty.target - Login Prompts. May 15 15:12:56.758958 containerd[1549]: time="2025-05-15T15:12:56.757398135Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="13.002µs" May 15 15:12:56.758958 containerd[1549]: time="2025-05-15T15:12:56.757438515Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 15 15:12:56.758958 containerd[1549]: time="2025-05-15T15:12:56.757457812Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 15 15:12:56.758958 containerd[1549]: time="2025-05-15T15:12:56.757666053Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 15 15:12:56.758958 containerd[1549]: time="2025-05-15T15:12:56.757689790Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 15 15:12:56.758958 containerd[1549]: time="2025-05-15T15:12:56.757716007Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 15 15:12:56.758958 containerd[1549]: time="2025-05-15T15:12:56.757770062Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 15 15:12:56.758958 containerd[1549]: time="2025-05-15T15:12:56.757781337Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 15 15:12:56.760363 containerd[1549]: time="2025-05-15T15:12:56.760329800Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 15 15:12:56.764321 kernel: EDAC MC: Ver: 3.0.0 May 15 15:12:56.764387 containerd[1549]: time="2025-05-15T15:12:56.763292860Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 15 15:12:56.764387 containerd[1549]: time="2025-05-15T15:12:56.763350136Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 15 15:12:56.764387 containerd[1549]: time="2025-05-15T15:12:56.763361457Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 15 15:12:56.764387 containerd[1549]: time="2025-05-15T15:12:56.763481458Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 15 15:12:56.764387 containerd[1549]: time="2025-05-15T15:12:56.763713569Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 15 15:12:56.764387 containerd[1549]: time="2025-05-15T15:12:56.763748683Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 15 15:12:56.764387 containerd[1549]: time="2025-05-15T15:12:56.763759063Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 15 15:12:56.765521 containerd[1549]: time="2025-05-15T15:12:56.765102545Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 15 15:12:56.771389 containerd[1549]: time="2025-05-15T15:12:56.770870851Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 15 15:12:56.771389 containerd[1549]: time="2025-05-15T15:12:56.771005493Z" level=info msg="metadata content store policy set" policy=shared May 15 15:12:56.774162 containerd[1549]: time="2025-05-15T15:12:56.774109207Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774361040Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774388303Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774438705Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774456024Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774471047Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774485863Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774515442Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774534588Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774547325Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774556865Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774569825Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774723040Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774749887Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 15 15:12:56.775363 containerd[1549]: time="2025-05-15T15:12:56.774764602Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774776205Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774787826Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774798037Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774811194Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774820833Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774831897Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774841462Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774851210Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 15 15:12:56.775705 containerd[1549]: time="2025-05-15T15:12:56.774914581Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 15 15:12:56.780346 containerd[1549]: time="2025-05-15T15:12:56.780277882Z" level=info msg="Start snapshots syncer" May 15 15:12:56.781959 containerd[1549]: time="2025-05-15T15:12:56.780473103Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 15 15:12:56.781959 containerd[1549]: time="2025-05-15T15:12:56.780847132Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 15 15:12:56.782249 containerd[1549]: time="2025-05-15T15:12:56.780904621Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 15 15:12:56.782249 containerd[1549]: time="2025-05-15T15:12:56.781119494Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782703910Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782755248Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782773103Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782789313Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782810700Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782826823Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782842017Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782879008Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782890517Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 15 15:12:56.783327 containerd[1549]: time="2025-05-15T15:12:56.782903361Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.783993728Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784401304Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784421268Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784437476Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784446181Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784457415Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784471913Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784489661Z" level=info msg="runtime interface created" May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784495244Z" level=info msg="created NRI interface" May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784508838Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784526671Z" level=info msg="Connect containerd service" May 15 15:12:56.784840 containerd[1549]: time="2025-05-15T15:12:56.784564462Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 15 15:12:56.789560 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 15 15:12:56.796471 containerd[1549]: time="2025-05-15T15:12:56.796006138Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 15 15:12:56.848672 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 15 15:12:56.849213 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 15 15:12:56.849593 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 15 15:12:56.852690 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 15 15:12:56.854678 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 15 15:12:56.960558 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 15 15:12:56.992162 containerd[1549]: time="2025-05-15T15:12:56.992112963Z" level=info msg="Start subscribing containerd event" May 15 15:12:56.992676 containerd[1549]: time="2025-05-15T15:12:56.992321066Z" level=info msg="Start recovering state" May 15 15:12:56.992676 containerd[1549]: time="2025-05-15T15:12:56.992487580Z" level=info msg="Start event monitor" May 15 15:12:56.992676 containerd[1549]: time="2025-05-15T15:12:56.992505594Z" level=info msg="Start cni network conf syncer for default" May 15 15:12:56.992676 containerd[1549]: time="2025-05-15T15:12:56.992513686Z" level=info msg="Start streaming server" May 15 15:12:56.992676 containerd[1549]: time="2025-05-15T15:12:56.992527172Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 15 15:12:56.992676 containerd[1549]: time="2025-05-15T15:12:56.992533555Z" level=info msg="runtime interface starting up..." May 15 15:12:56.992676 containerd[1549]: time="2025-05-15T15:12:56.992538624Z" level=info msg="starting plugins..." May 15 15:12:56.992676 containerd[1549]: time="2025-05-15T15:12:56.992551255Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 15 15:12:56.993091 containerd[1549]: time="2025-05-15T15:12:56.992989891Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 15 15:12:56.993694 containerd[1549]: time="2025-05-15T15:12:56.993660175Z" level=info msg=serving... address=/run/containerd/containerd.sock May 15 15:12:56.994408 containerd[1549]: time="2025-05-15T15:12:56.994306019Z" level=info msg="containerd successfully booted in 0.378978s" May 15 15:12:56.994527 systemd[1]: Started containerd.service - containerd container runtime. May 15 15:12:57.167085 systemd-networkd[1455]: eth0: Gained IPv6LL May 15 15:12:57.171899 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 15 15:12:57.172499 systemd[1]: Reached target network-online.target - Network is Online. May 15 15:12:57.175247 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 15:12:57.178180 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 15 15:12:57.230691 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 15 15:12:57.233091 systemd-networkd[1455]: eth1: Gained IPv6LL May 15 15:12:57.236970 tar[1538]: linux-amd64/README.md May 15 15:12:57.251966 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 15 15:12:58.224294 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 15 15:12:58.224904 systemd[1]: Reached target multi-user.target - Multi-User System. May 15 15:12:58.225825 systemd[1]: Startup finished in 4.402s (kernel) + 5.668s (initrd) + 5.579s (userspace) = 15.650s. May 15 15:12:58.230481 (kubelet)[1684]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 15 15:12:58.789743 kubelet[1684]: E0515 15:12:58.789644 1684 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 15 15:12:58.792178 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 15 15:12:58.792354 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 15 15:12:58.793096 systemd[1]: kubelet.service: Consumed 1.189s CPU time, 252.5M memory peak. May 15 15:13:01.164678 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 15 15:13:01.167241 systemd[1]: Started sshd@0-24.144.90.91:22-139.178.68.195:57628.service - OpenSSH per-connection server daemon (139.178.68.195:57628). May 15 15:13:01.264296 sshd[1696]: Accepted publickey for core from 139.178.68.195 port 57628 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:13:01.266737 sshd-session[1696]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:13:01.277808 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 15 15:13:01.279121 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 15 15:13:01.290874 systemd-logind[1527]: New session 1 of user core. May 15 15:13:01.308540 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 15 15:13:01.314357 systemd[1]: Starting user@500.service - User Manager for UID 500... May 15 15:13:01.333620 (systemd)[1700]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 15 15:13:01.337525 systemd-logind[1527]: New session c1 of user core. May 15 15:13:01.554193 systemd[1700]: Queued start job for default target default.target. May 15 15:13:01.564716 systemd[1700]: Created slice app.slice - User Application Slice. May 15 15:13:01.564778 systemd[1700]: Reached target paths.target - Paths. May 15 15:13:01.564846 systemd[1700]: Reached target timers.target - Timers. May 15 15:13:01.567048 systemd[1700]: Starting dbus.socket - D-Bus User Message Bus Socket... May 15 15:13:01.585456 systemd[1700]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 15 15:13:01.585847 systemd[1700]: Reached target sockets.target - Sockets. May 15 15:13:01.586034 systemd[1700]: Reached target basic.target - Basic System. May 15 15:13:01.586192 systemd[1700]: Reached target default.target - Main User Target. May 15 15:13:01.586226 systemd[1]: Started user@500.service - User Manager for UID 500. May 15 15:13:01.586407 systemd[1700]: Startup finished in 240ms. May 15 15:13:01.598830 systemd[1]: Started session-1.scope - Session 1 of User core. May 15 15:13:01.671306 systemd[1]: Started sshd@1-24.144.90.91:22-139.178.68.195:57630.service - OpenSSH per-connection server daemon (139.178.68.195:57630). May 15 15:13:01.732973 sshd[1711]: Accepted publickey for core from 139.178.68.195 port 57630 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:13:01.735100 sshd-session[1711]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:13:01.742450 systemd-logind[1527]: New session 2 of user core. May 15 15:13:01.748280 systemd[1]: Started session-2.scope - Session 2 of User core. May 15 15:13:01.817897 sshd[1713]: Connection closed by 139.178.68.195 port 57630 May 15 15:13:01.819029 sshd-session[1711]: pam_unix(sshd:session): session closed for user core May 15 15:13:01.829637 systemd[1]: sshd@1-24.144.90.91:22-139.178.68.195:57630.service: Deactivated successfully. May 15 15:13:01.832033 systemd[1]: session-2.scope: Deactivated successfully. May 15 15:13:01.833502 systemd-logind[1527]: Session 2 logged out. Waiting for processes to exit. May 15 15:13:01.838243 systemd[1]: Started sshd@2-24.144.90.91:22-139.178.68.195:57642.service - OpenSSH per-connection server daemon (139.178.68.195:57642). May 15 15:13:01.840031 systemd-logind[1527]: Removed session 2. May 15 15:13:01.908680 sshd[1719]: Accepted publickey for core from 139.178.68.195 port 57642 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:13:01.910581 sshd-session[1719]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:13:01.916581 systemd-logind[1527]: New session 3 of user core. May 15 15:13:01.924310 systemd[1]: Started session-3.scope - Session 3 of User core. May 15 15:13:01.982703 sshd[1721]: Connection closed by 139.178.68.195 port 57642 May 15 15:13:01.983517 sshd-session[1719]: pam_unix(sshd:session): session closed for user core May 15 15:13:02.000457 systemd[1]: sshd@2-24.144.90.91:22-139.178.68.195:57642.service: Deactivated successfully. May 15 15:13:02.003070 systemd[1]: session-3.scope: Deactivated successfully. May 15 15:13:02.004361 systemd-logind[1527]: Session 3 logged out. Waiting for processes to exit. May 15 15:13:02.011393 systemd[1]: Started sshd@3-24.144.90.91:22-139.178.68.195:57648.service - OpenSSH per-connection server daemon (139.178.68.195:57648). May 15 15:13:02.012903 systemd-logind[1527]: Removed session 3. May 15 15:13:02.078740 sshd[1727]: Accepted publickey for core from 139.178.68.195 port 57648 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:13:02.082345 sshd-session[1727]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:13:02.091085 systemd-logind[1527]: New session 4 of user core. May 15 15:13:02.104263 systemd[1]: Started session-4.scope - Session 4 of User core. May 15 15:13:02.168441 sshd[1729]: Connection closed by 139.178.68.195 port 57648 May 15 15:13:02.169219 sshd-session[1727]: pam_unix(sshd:session): session closed for user core May 15 15:13:02.181803 systemd[1]: sshd@3-24.144.90.91:22-139.178.68.195:57648.service: Deactivated successfully. May 15 15:13:02.184982 systemd[1]: session-4.scope: Deactivated successfully. May 15 15:13:02.186353 systemd-logind[1527]: Session 4 logged out. Waiting for processes to exit. May 15 15:13:02.190846 systemd[1]: Started sshd@4-24.144.90.91:22-139.178.68.195:57654.service - OpenSSH per-connection server daemon (139.178.68.195:57654). May 15 15:13:02.192243 systemd-logind[1527]: Removed session 4. May 15 15:13:02.258326 sshd[1735]: Accepted publickey for core from 139.178.68.195 port 57654 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:13:02.260326 sshd-session[1735]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:13:02.267005 systemd-logind[1527]: New session 5 of user core. May 15 15:13:02.276258 systemd[1]: Started session-5.scope - Session 5 of User core. May 15 15:13:02.359249 sudo[1738]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 15 15:13:02.359679 sudo[1738]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 15 15:13:02.373378 sudo[1738]: pam_unix(sudo:session): session closed for user root May 15 15:13:02.377966 sshd[1737]: Connection closed by 139.178.68.195 port 57654 May 15 15:13:02.377732 sshd-session[1735]: pam_unix(sshd:session): session closed for user core May 15 15:13:02.394595 systemd[1]: sshd@4-24.144.90.91:22-139.178.68.195:57654.service: Deactivated successfully. May 15 15:13:02.397651 systemd[1]: session-5.scope: Deactivated successfully. May 15 15:13:02.399457 systemd-logind[1527]: Session 5 logged out. Waiting for processes to exit. May 15 15:13:02.404290 systemd[1]: Started sshd@5-24.144.90.91:22-139.178.68.195:57656.service - OpenSSH per-connection server daemon (139.178.68.195:57656). May 15 15:13:02.405789 systemd-logind[1527]: Removed session 5. May 15 15:13:02.463250 sshd[1744]: Accepted publickey for core from 139.178.68.195 port 57656 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:13:02.465127 sshd-session[1744]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:13:02.473024 systemd-logind[1527]: New session 6 of user core. May 15 15:13:02.481793 systemd[1]: Started session-6.scope - Session 6 of User core. May 15 15:13:02.543631 sudo[1748]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 15 15:13:02.544571 sudo[1748]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 15 15:13:02.618207 sudo[1748]: pam_unix(sudo:session): session closed for user root May 15 15:13:02.628995 sudo[1747]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 15 15:13:02.629521 sudo[1747]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 15 15:13:02.645033 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 15 15:13:02.700176 augenrules[1770]: No rules May 15 15:13:02.701836 systemd[1]: audit-rules.service: Deactivated successfully. May 15 15:13:02.702250 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 15 15:13:02.703809 sudo[1747]: pam_unix(sudo:session): session closed for user root May 15 15:13:02.707587 sshd[1746]: Connection closed by 139.178.68.195 port 57656 May 15 15:13:02.708456 sshd-session[1744]: pam_unix(sshd:session): session closed for user core May 15 15:13:02.728041 systemd[1]: sshd@5-24.144.90.91:22-139.178.68.195:57656.service: Deactivated successfully. May 15 15:13:02.731026 systemd[1]: session-6.scope: Deactivated successfully. May 15 15:13:02.732616 systemd-logind[1527]: Session 6 logged out. Waiting for processes to exit. May 15 15:13:02.736874 systemd[1]: Started sshd@6-24.144.90.91:22-139.178.68.195:57670.service - OpenSSH per-connection server daemon (139.178.68.195:57670). May 15 15:13:02.739042 systemd-logind[1527]: Removed session 6. May 15 15:13:02.793966 sshd[1779]: Accepted publickey for core from 139.178.68.195 port 57670 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:13:02.796043 sshd-session[1779]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:13:02.803545 systemd-logind[1527]: New session 7 of user core. May 15 15:13:02.814341 systemd[1]: Started session-7.scope - Session 7 of User core. May 15 15:13:02.875889 sudo[1782]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 15 15:13:02.877115 sudo[1782]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 15 15:13:03.460809 systemd[1]: Starting docker.service - Docker Application Container Engine... May 15 15:13:03.475824 (dockerd)[1799]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 15 15:13:03.854484 dockerd[1799]: time="2025-05-15T15:13:03.854019706Z" level=info msg="Starting up" May 15 15:13:03.855390 dockerd[1799]: time="2025-05-15T15:13:03.855343745Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 15 15:13:03.926511 dockerd[1799]: time="2025-05-15T15:13:03.926437923Z" level=info msg="Loading containers: start." May 15 15:13:03.940026 kernel: Initializing XFRM netlink socket May 15 15:13:04.264742 systemd-networkd[1455]: docker0: Link UP May 15 15:13:04.268846 dockerd[1799]: time="2025-05-15T15:13:04.268760013Z" level=info msg="Loading containers: done." May 15 15:13:04.288475 dockerd[1799]: time="2025-05-15T15:13:04.288397933Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 15 15:13:04.288717 dockerd[1799]: time="2025-05-15T15:13:04.288539909Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 May 15 15:13:04.288775 dockerd[1799]: time="2025-05-15T15:13:04.288715717Z" level=info msg="Initializing buildkit" May 15 15:13:04.317152 dockerd[1799]: time="2025-05-15T15:13:04.316993916Z" level=info msg="Completed buildkit initialization" May 15 15:13:04.330225 dockerd[1799]: time="2025-05-15T15:13:04.329511309Z" level=info msg="Daemon has completed initialization" May 15 15:13:04.329866 systemd[1]: Started docker.service - Docker Application Container Engine. May 15 15:13:04.330542 dockerd[1799]: time="2025-05-15T15:13:04.330497354Z" level=info msg="API listen on /run/docker.sock" May 15 15:13:05.270231 containerd[1549]: time="2025-05-15T15:13:05.270180688Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\"" May 15 15:13:05.875766 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1578649762.mount: Deactivated successfully. May 15 15:13:07.202732 containerd[1549]: time="2025-05-15T15:13:07.202661480Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:07.203848 containerd[1549]: time="2025-05-15T15:13:07.203725915Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.4: active requests=0, bytes read=28682879" May 15 15:13:07.204379 containerd[1549]: time="2025-05-15T15:13:07.204343329Z" level=info msg="ImageCreate event name:\"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:07.206896 containerd[1549]: time="2025-05-15T15:13:07.206857165Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:07.208542 containerd[1549]: time="2025-05-15T15:13:07.208485230Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.4\" with image id \"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\", size \"28679679\" in 1.938247804s" May 15 15:13:07.208943 containerd[1549]: time="2025-05-15T15:13:07.208696569Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\" returns image reference \"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\"" May 15 15:13:07.209500 containerd[1549]: time="2025-05-15T15:13:07.209462771Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\"" May 15 15:13:08.687085 containerd[1549]: time="2025-05-15T15:13:08.686910040Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:08.689138 containerd[1549]: time="2025-05-15T15:13:08.689050187Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.4: active requests=0, bytes read=24779589" May 15 15:13:08.689731 containerd[1549]: time="2025-05-15T15:13:08.689675948Z" level=info msg="ImageCreate event name:\"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:08.693108 containerd[1549]: time="2025-05-15T15:13:08.693040144Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:08.695144 containerd[1549]: time="2025-05-15T15:13:08.694784609Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.4\" with image id \"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\", size \"26267962\" in 1.485292866s" May 15 15:13:08.695144 containerd[1549]: time="2025-05-15T15:13:08.694827196Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\" returns image reference \"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\"" May 15 15:13:08.695571 containerd[1549]: time="2025-05-15T15:13:08.695547186Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\"" May 15 15:13:08.930647 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 15 15:13:08.932369 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 15:13:09.116409 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 15 15:13:09.129533 (kubelet)[2076]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 15 15:13:09.183832 kubelet[2076]: E0515 15:13:09.183734 2076 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 15 15:13:09.187797 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 15 15:13:09.187978 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 15 15:13:09.188496 systemd[1]: kubelet.service: Consumed 205ms CPU time, 102.6M memory peak. May 15 15:13:10.058986 containerd[1549]: time="2025-05-15T15:13:10.058458172Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:10.059460 containerd[1549]: time="2025-05-15T15:13:10.059391955Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.4: active requests=0, bytes read=19169938" May 15 15:13:10.059877 containerd[1549]: time="2025-05-15T15:13:10.059837320Z" level=info msg="ImageCreate event name:\"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:10.063500 containerd[1549]: time="2025-05-15T15:13:10.063441093Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:10.064796 containerd[1549]: time="2025-05-15T15:13:10.064611778Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.4\" with image id \"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\", size \"20658329\" in 1.369030389s" May 15 15:13:10.064796 containerd[1549]: time="2025-05-15T15:13:10.064663984Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\" returns image reference \"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\"" May 15 15:13:10.065532 containerd[1549]: time="2025-05-15T15:13:10.065396540Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\"" May 15 15:13:11.251734 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2221224702.mount: Deactivated successfully. May 15 15:13:11.782745 containerd[1549]: time="2025-05-15T15:13:11.782690006Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:11.783640 containerd[1549]: time="2025-05-15T15:13:11.783378883Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.4: active requests=0, bytes read=30917856" May 15 15:13:11.784257 containerd[1549]: time="2025-05-15T15:13:11.784216885Z" level=info msg="ImageCreate event name:\"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:11.787161 containerd[1549]: time="2025-05-15T15:13:11.787116686Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:11.788038 containerd[1549]: time="2025-05-15T15:13:11.787999249Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.4\" with image id \"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\", repo tag \"registry.k8s.io/kube-proxy:v1.32.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\", size \"30916875\" in 1.722303273s" May 15 15:13:11.788281 containerd[1549]: time="2025-05-15T15:13:11.788162593Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\" returns image reference \"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\"" May 15 15:13:11.789280 containerd[1549]: time="2025-05-15T15:13:11.789094816Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" May 15 15:13:11.790533 systemd-resolved[1408]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. May 15 15:13:12.233920 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2973495863.mount: Deactivated successfully. May 15 15:13:13.176991 containerd[1549]: time="2025-05-15T15:13:13.176803348Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:13.178744 containerd[1549]: time="2025-05-15T15:13:13.178678038Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" May 15 15:13:13.179573 containerd[1549]: time="2025-05-15T15:13:13.179522344Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:13.182591 containerd[1549]: time="2025-05-15T15:13:13.182479251Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:13.184542 containerd[1549]: time="2025-05-15T15:13:13.184066901Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.394919461s" May 15 15:13:13.184542 containerd[1549]: time="2025-05-15T15:13:13.184123181Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" May 15 15:13:13.184875 containerd[1549]: time="2025-05-15T15:13:13.184842897Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" May 15 15:13:13.585031 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1387561853.mount: Deactivated successfully. May 15 15:13:13.589486 containerd[1549]: time="2025-05-15T15:13:13.588429366Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 15 15:13:13.589486 containerd[1549]: time="2025-05-15T15:13:13.589352067Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" May 15 15:13:13.589486 containerd[1549]: time="2025-05-15T15:13:13.589420783Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 15 15:13:13.591382 containerd[1549]: time="2025-05-15T15:13:13.591337086Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 15 15:13:13.592114 containerd[1549]: time="2025-05-15T15:13:13.592082561Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 407.205266ms" May 15 15:13:13.592114 containerd[1549]: time="2025-05-15T15:13:13.592117308Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" May 15 15:13:13.592815 containerd[1549]: time="2025-05-15T15:13:13.592789615Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" May 15 15:13:14.063424 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3361549070.mount: Deactivated successfully. May 15 15:13:14.895195 systemd-resolved[1408]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. May 15 15:13:15.813239 containerd[1549]: time="2025-05-15T15:13:15.812139486Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:15.813239 containerd[1549]: time="2025-05-15T15:13:15.812966239Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57551360" May 15 15:13:15.813706 containerd[1549]: time="2025-05-15T15:13:15.813645211Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:15.817537 containerd[1549]: time="2025-05-15T15:13:15.817476105Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:15.819086 containerd[1549]: time="2025-05-15T15:13:15.819026293Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.226203421s" May 15 15:13:15.819086 containerd[1549]: time="2025-05-15T15:13:15.819087651Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" May 15 15:13:18.854637 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 15 15:13:18.855357 systemd[1]: kubelet.service: Consumed 205ms CPU time, 102.6M memory peak. May 15 15:13:18.859538 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 15:13:18.897976 systemd[1]: Reload requested from client PID 2234 ('systemctl') (unit session-7.scope)... May 15 15:13:18.898003 systemd[1]: Reloading... May 15 15:13:19.032020 zram_generator::config[2277]: No configuration found. May 15 15:13:19.139792 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 15 15:13:19.268414 systemd[1]: Reloading finished in 369 ms. May 15 15:13:19.346986 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM May 15 15:13:19.347130 systemd[1]: kubelet.service: Failed with result 'signal'. May 15 15:13:19.347531 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 15 15:13:19.347610 systemd[1]: kubelet.service: Consumed 106ms CPU time, 91.8M memory peak. May 15 15:13:19.350157 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 15:13:19.502676 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 15 15:13:19.516459 (kubelet)[2331]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 15 15:13:19.563265 kubelet[2331]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 15 15:13:19.563988 kubelet[2331]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 15 15:13:19.563988 kubelet[2331]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 15 15:13:19.564167 kubelet[2331]: I0515 15:13:19.564123 2331 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 15 15:13:20.078859 kubelet[2331]: I0515 15:13:20.078100 2331 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" May 15 15:13:20.078859 kubelet[2331]: I0515 15:13:20.078156 2331 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 15 15:13:20.079450 kubelet[2331]: I0515 15:13:20.079287 2331 server.go:954] "Client rotation is on, will bootstrap in background" May 15 15:13:20.103017 kubelet[2331]: I0515 15:13:20.102826 2331 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 15 15:13:20.106101 kubelet[2331]: E0515 15:13:20.106040 2331 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://24.144.90.91:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 24.144.90.91:6443: connect: connection refused" logger="UnhandledError" May 15 15:13:20.114538 kubelet[2331]: I0515 15:13:20.114492 2331 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 15 15:13:20.119489 kubelet[2331]: I0515 15:13:20.119446 2331 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 15 15:13:20.120719 kubelet[2331]: I0515 15:13:20.120664 2331 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 15 15:13:20.120894 kubelet[2331]: I0515 15:13:20.120715 2331 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4334.0.0-a-073c24674a","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 15 15:13:20.121042 kubelet[2331]: I0515 15:13:20.120898 2331 topology_manager.go:138] "Creating topology manager with none policy" May 15 15:13:20.121042 kubelet[2331]: I0515 15:13:20.120909 2331 container_manager_linux.go:304] "Creating device plugin manager" May 15 15:13:20.121108 kubelet[2331]: I0515 15:13:20.121047 2331 state_mem.go:36] "Initialized new in-memory state store" May 15 15:13:20.124141 kubelet[2331]: I0515 15:13:20.124097 2331 kubelet.go:446] "Attempting to sync node with API server" May 15 15:13:20.124251 kubelet[2331]: I0515 15:13:20.124210 2331 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 15 15:13:20.124289 kubelet[2331]: I0515 15:13:20.124255 2331 kubelet.go:352] "Adding apiserver pod source" May 15 15:13:20.124289 kubelet[2331]: I0515 15:13:20.124274 2331 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 15 15:13:20.130760 kubelet[2331]: W0515 15:13:20.129975 2331 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://24.144.90.91:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4334.0.0-a-073c24674a&limit=500&resourceVersion=0": dial tcp 24.144.90.91:6443: connect: connection refused May 15 15:13:20.130760 kubelet[2331]: E0515 15:13:20.130040 2331 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://24.144.90.91:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4334.0.0-a-073c24674a&limit=500&resourceVersion=0\": dial tcp 24.144.90.91:6443: connect: connection refused" logger="UnhandledError" May 15 15:13:20.130760 kubelet[2331]: W0515 15:13:20.130539 2331 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://24.144.90.91:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 24.144.90.91:6443: connect: connection refused May 15 15:13:20.130760 kubelet[2331]: E0515 15:13:20.130576 2331 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://24.144.90.91:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 24.144.90.91:6443: connect: connection refused" logger="UnhandledError" May 15 15:13:20.132258 kubelet[2331]: I0515 15:13:20.132223 2331 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 15 15:13:20.137413 kubelet[2331]: I0515 15:13:20.137374 2331 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 15 15:13:20.138232 kubelet[2331]: W0515 15:13:20.138117 2331 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 15 15:13:20.141975 kubelet[2331]: I0515 15:13:20.141021 2331 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 15 15:13:20.141975 kubelet[2331]: I0515 15:13:20.141062 2331 server.go:1287] "Started kubelet" May 15 15:13:20.141975 kubelet[2331]: I0515 15:13:20.141404 2331 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 15 15:13:20.142774 kubelet[2331]: I0515 15:13:20.142280 2331 server.go:490] "Adding debug handlers to kubelet server" May 15 15:13:20.143495 kubelet[2331]: I0515 15:13:20.143191 2331 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 15 15:13:20.143495 kubelet[2331]: I0515 15:13:20.143437 2331 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 15 15:13:20.145064 kubelet[2331]: I0515 15:13:20.145041 2331 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 15 15:13:20.152697 kubelet[2331]: E0515 15:13:20.148950 2331 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://24.144.90.91:6443/api/v1/namespaces/default/events\": dial tcp 24.144.90.91:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4334.0.0-a-073c24674a.183fbc19f6a1bf87 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4334.0.0-a-073c24674a,UID:ci-4334.0.0-a-073c24674a,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4334.0.0-a-073c24674a,},FirstTimestamp:2025-05-15 15:13:20.141041543 +0000 UTC m=+0.618953120,LastTimestamp:2025-05-15 15:13:20.141041543 +0000 UTC m=+0.618953120,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4334.0.0-a-073c24674a,}" May 15 15:13:20.154420 kubelet[2331]: E0515 15:13:20.154392 2331 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 15 15:13:20.154588 kubelet[2331]: I0515 15:13:20.154572 2331 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 15 15:13:20.157122 kubelet[2331]: E0515 15:13:20.157087 2331 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4334.0.0-a-073c24674a\" not found" May 15 15:13:20.157251 kubelet[2331]: I0515 15:13:20.157130 2331 volume_manager.go:297] "Starting Kubelet Volume Manager" May 15 15:13:20.157357 kubelet[2331]: I0515 15:13:20.157342 2331 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 15 15:13:20.157424 kubelet[2331]: I0515 15:13:20.157414 2331 reconciler.go:26] "Reconciler: start to sync state" May 15 15:13:20.158072 kubelet[2331]: W0515 15:13:20.158031 2331 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://24.144.90.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 24.144.90.91:6443: connect: connection refused May 15 15:13:20.158142 kubelet[2331]: E0515 15:13:20.158092 2331 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://24.144.90.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 24.144.90.91:6443: connect: connection refused" logger="UnhandledError" May 15 15:13:20.158599 kubelet[2331]: E0515 15:13:20.158566 2331 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.144.90.91:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4334.0.0-a-073c24674a?timeout=10s\": dial tcp 24.144.90.91:6443: connect: connection refused" interval="200ms" May 15 15:13:20.159106 kubelet[2331]: I0515 15:13:20.159081 2331 factory.go:221] Registration of the systemd container factory successfully May 15 15:13:20.159223 kubelet[2331]: I0515 15:13:20.159201 2331 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 15 15:13:20.162486 kubelet[2331]: I0515 15:13:20.162450 2331 factory.go:221] Registration of the containerd container factory successfully May 15 15:13:20.175468 kubelet[2331]: I0515 15:13:20.175128 2331 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 15 15:13:20.177152 kubelet[2331]: I0515 15:13:20.177113 2331 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 15 15:13:20.177152 kubelet[2331]: I0515 15:13:20.177153 2331 status_manager.go:227] "Starting to sync pod status with apiserver" May 15 15:13:20.177288 kubelet[2331]: I0515 15:13:20.177185 2331 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 15 15:13:20.177288 kubelet[2331]: I0515 15:13:20.177241 2331 kubelet.go:2388] "Starting kubelet main sync loop" May 15 15:13:20.177347 kubelet[2331]: E0515 15:13:20.177320 2331 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 15 15:13:20.185854 kubelet[2331]: W0515 15:13:20.185771 2331 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://24.144.90.91:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 24.144.90.91:6443: connect: connection refused May 15 15:13:20.186078 kubelet[2331]: E0515 15:13:20.185869 2331 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://24.144.90.91:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 24.144.90.91:6443: connect: connection refused" logger="UnhandledError" May 15 15:13:20.195471 kubelet[2331]: I0515 15:13:20.195424 2331 cpu_manager.go:221] "Starting CPU manager" policy="none" May 15 15:13:20.195471 kubelet[2331]: I0515 15:13:20.195449 2331 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 15 15:13:20.195471 kubelet[2331]: I0515 15:13:20.195474 2331 state_mem.go:36] "Initialized new in-memory state store" May 15 15:13:20.196951 kubelet[2331]: I0515 15:13:20.196890 2331 policy_none.go:49] "None policy: Start" May 15 15:13:20.196951 kubelet[2331]: I0515 15:13:20.196923 2331 memory_manager.go:186] "Starting memorymanager" policy="None" May 15 15:13:20.197147 kubelet[2331]: I0515 15:13:20.196971 2331 state_mem.go:35] "Initializing new in-memory state store" May 15 15:13:20.207748 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 15 15:13:20.221088 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 15 15:13:20.257566 kubelet[2331]: E0515 15:13:20.257412 2331 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4334.0.0-a-073c24674a\" not found" May 15 15:13:20.262404 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 15 15:13:20.277543 kubelet[2331]: E0515 15:13:20.277487 2331 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 15 15:13:20.283358 kubelet[2331]: I0515 15:13:20.283322 2331 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 15 15:13:20.285998 kubelet[2331]: I0515 15:13:20.285737 2331 eviction_manager.go:189] "Eviction manager: starting control loop" May 15 15:13:20.285998 kubelet[2331]: I0515 15:13:20.285793 2331 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 15 15:13:20.286959 kubelet[2331]: I0515 15:13:20.286699 2331 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 15 15:13:20.288960 kubelet[2331]: E0515 15:13:20.288300 2331 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 15 15:13:20.289069 kubelet[2331]: E0515 15:13:20.288988 2331 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4334.0.0-a-073c24674a\" not found" May 15 15:13:20.359584 kubelet[2331]: E0515 15:13:20.359422 2331 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.144.90.91:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4334.0.0-a-073c24674a?timeout=10s\": dial tcp 24.144.90.91:6443: connect: connection refused" interval="400ms" May 15 15:13:20.387650 kubelet[2331]: I0515 15:13:20.387587 2331 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-073c24674a" May 15 15:13:20.388467 kubelet[2331]: E0515 15:13:20.388413 2331 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://24.144.90.91:6443/api/v1/nodes\": dial tcp 24.144.90.91:6443: connect: connection refused" node="ci-4334.0.0-a-073c24674a" May 15 15:13:20.489552 systemd[1]: Created slice kubepods-burstable-podb49fd5be0a8bc73cf8e412ce96b1971d.slice - libcontainer container kubepods-burstable-podb49fd5be0a8bc73cf8e412ce96b1971d.slice. May 15 15:13:20.509300 kubelet[2331]: E0515 15:13:20.508155 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:20.511894 systemd[1]: Created slice kubepods-burstable-pod0e261e850a4cc2912c11fa835e595fef.slice - libcontainer container kubepods-burstable-pod0e261e850a4cc2912c11fa835e595fef.slice. May 15 15:13:20.520624 systemd[1]: Created slice kubepods-burstable-podb373e34f35214e049bae3b4082d9f52b.slice - libcontainer container kubepods-burstable-podb373e34f35214e049bae3b4082d9f52b.slice. May 15 15:13:20.523878 kubelet[2331]: E0515 15:13:20.523831 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:20.524135 kubelet[2331]: E0515 15:13:20.524113 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:20.559109 kubelet[2331]: I0515 15:13:20.559051 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-k8s-certs\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:20.559475 kubelet[2331]: I0515 15:13:20.559418 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-kubeconfig\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:20.559647 kubelet[2331]: I0515 15:13:20.559576 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-flexvolume-dir\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:20.559794 kubelet[2331]: I0515 15:13:20.559771 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:20.560001 kubelet[2331]: I0515 15:13:20.559924 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0e261e850a4cc2912c11fa835e595fef-kubeconfig\") pod \"kube-scheduler-ci-4334.0.0-a-073c24674a\" (UID: \"0e261e850a4cc2912c11fa835e595fef\") " pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:20.560192 kubelet[2331]: I0515 15:13:20.560146 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b49fd5be0a8bc73cf8e412ce96b1971d-ca-certs\") pod \"kube-apiserver-ci-4334.0.0-a-073c24674a\" (UID: \"b49fd5be0a8bc73cf8e412ce96b1971d\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:20.560316 kubelet[2331]: I0515 15:13:20.560299 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b49fd5be0a8bc73cf8e412ce96b1971d-k8s-certs\") pod \"kube-apiserver-ci-4334.0.0-a-073c24674a\" (UID: \"b49fd5be0a8bc73cf8e412ce96b1971d\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:20.560521 kubelet[2331]: I0515 15:13:20.560472 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b49fd5be0a8bc73cf8e412ce96b1971d-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4334.0.0-a-073c24674a\" (UID: \"b49fd5be0a8bc73cf8e412ce96b1971d\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:20.560630 kubelet[2331]: I0515 15:13:20.560501 2331 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-ca-certs\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:20.590854 kubelet[2331]: I0515 15:13:20.590477 2331 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-073c24674a" May 15 15:13:20.591340 kubelet[2331]: E0515 15:13:20.590946 2331 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://24.144.90.91:6443/api/v1/nodes\": dial tcp 24.144.90.91:6443: connect: connection refused" node="ci-4334.0.0-a-073c24674a" May 15 15:13:20.760559 kubelet[2331]: E0515 15:13:20.760490 2331 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.144.90.91:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4334.0.0-a-073c24674a?timeout=10s\": dial tcp 24.144.90.91:6443: connect: connection refused" interval="800ms" May 15 15:13:20.809049 kubelet[2331]: E0515 15:13:20.808996 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:20.809745 containerd[1549]: time="2025-05-15T15:13:20.809708104Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4334.0.0-a-073c24674a,Uid:b49fd5be0a8bc73cf8e412ce96b1971d,Namespace:kube-system,Attempt:0,}" May 15 15:13:20.826518 kubelet[2331]: E0515 15:13:20.826178 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:20.827009 kubelet[2331]: E0515 15:13:20.826987 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:20.833745 containerd[1549]: time="2025-05-15T15:13:20.833395976Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4334.0.0-a-073c24674a,Uid:0e261e850a4cc2912c11fa835e595fef,Namespace:kube-system,Attempt:0,}" May 15 15:13:20.834505 containerd[1549]: time="2025-05-15T15:13:20.834465214Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4334.0.0-a-073c24674a,Uid:b373e34f35214e049bae3b4082d9f52b,Namespace:kube-system,Attempt:0,}" May 15 15:13:20.919808 containerd[1549]: time="2025-05-15T15:13:20.919723905Z" level=info msg="connecting to shim 900e02330d6a2d536094666bd2e5bece804162c265a360541832226b9718c8bd" address="unix:///run/containerd/s/b020c4387b1eb0b29cd54c0af9394ef9030b10ba013c7c6922ede5c97707e8dc" namespace=k8s.io protocol=ttrpc version=3 May 15 15:13:20.921097 containerd[1549]: time="2025-05-15T15:13:20.921036900Z" level=info msg="connecting to shim 901f146dfd565dc6c045a26e3f0f2d7f87f7d8e788c777f3c354cb8a8d1cb459" address="unix:///run/containerd/s/ecf854393674adc0ec17687ae5ae0c0f405df383c566396e63200a7018a88ff7" namespace=k8s.io protocol=ttrpc version=3 May 15 15:13:20.924128 containerd[1549]: time="2025-05-15T15:13:20.924062396Z" level=info msg="connecting to shim 0b30890a417013c759b59380b1d9a38e1d8f4d53c2ac152686e21a83c7bc1b1d" address="unix:///run/containerd/s/8fe0f3a5d001f8ffa0c7bf145154e261427abb6c7459f3bfa6cad2aed2f7982e" namespace=k8s.io protocol=ttrpc version=3 May 15 15:13:20.993583 kubelet[2331]: I0515 15:13:20.993281 2331 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-073c24674a" May 15 15:13:20.993742 kubelet[2331]: E0515 15:13:20.993649 2331 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://24.144.90.91:6443/api/v1/nodes\": dial tcp 24.144.90.91:6443: connect: connection refused" node="ci-4334.0.0-a-073c24674a" May 15 15:13:21.002463 kubelet[2331]: W0515 15:13:21.002279 2331 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://24.144.90.91:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 24.144.90.91:6443: connect: connection refused May 15 15:13:21.002463 kubelet[2331]: E0515 15:13:21.002348 2331 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://24.144.90.91:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 24.144.90.91:6443: connect: connection refused" logger="UnhandledError" May 15 15:13:21.014113 systemd[1]: Started cri-containerd-901f146dfd565dc6c045a26e3f0f2d7f87f7d8e788c777f3c354cb8a8d1cb459.scope - libcontainer container 901f146dfd565dc6c045a26e3f0f2d7f87f7d8e788c777f3c354cb8a8d1cb459. May 15 15:13:21.021744 systemd[1]: Started cri-containerd-0b30890a417013c759b59380b1d9a38e1d8f4d53c2ac152686e21a83c7bc1b1d.scope - libcontainer container 0b30890a417013c759b59380b1d9a38e1d8f4d53c2ac152686e21a83c7bc1b1d. May 15 15:13:21.024245 systemd[1]: Started cri-containerd-900e02330d6a2d536094666bd2e5bece804162c265a360541832226b9718c8bd.scope - libcontainer container 900e02330d6a2d536094666bd2e5bece804162c265a360541832226b9718c8bd. May 15 15:13:21.098694 containerd[1549]: time="2025-05-15T15:13:21.098577696Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4334.0.0-a-073c24674a,Uid:b49fd5be0a8bc73cf8e412ce96b1971d,Namespace:kube-system,Attempt:0,} returns sandbox id \"901f146dfd565dc6c045a26e3f0f2d7f87f7d8e788c777f3c354cb8a8d1cb459\"" May 15 15:13:21.101580 kubelet[2331]: E0515 15:13:21.101541 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:21.108252 containerd[1549]: time="2025-05-15T15:13:21.108200591Z" level=info msg="CreateContainer within sandbox \"901f146dfd565dc6c045a26e3f0f2d7f87f7d8e788c777f3c354cb8a8d1cb459\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 15 15:13:21.125720 containerd[1549]: time="2025-05-15T15:13:21.125168317Z" level=info msg="Container f4e736525f881446a5987a4aedead0e1a1abf56b7809b2a3bd34a76ed8bcbc85: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:21.135421 containerd[1549]: time="2025-05-15T15:13:21.135381134Z" level=info msg="CreateContainer within sandbox \"901f146dfd565dc6c045a26e3f0f2d7f87f7d8e788c777f3c354cb8a8d1cb459\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"f4e736525f881446a5987a4aedead0e1a1abf56b7809b2a3bd34a76ed8bcbc85\"" May 15 15:13:21.136864 containerd[1549]: time="2025-05-15T15:13:21.136801448Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4334.0.0-a-073c24674a,Uid:0e261e850a4cc2912c11fa835e595fef,Namespace:kube-system,Attempt:0,} returns sandbox id \"0b30890a417013c759b59380b1d9a38e1d8f4d53c2ac152686e21a83c7bc1b1d\"" May 15 15:13:21.137336 containerd[1549]: time="2025-05-15T15:13:21.137309911Z" level=info msg="StartContainer for \"f4e736525f881446a5987a4aedead0e1a1abf56b7809b2a3bd34a76ed8bcbc85\"" May 15 15:13:21.137820 kubelet[2331]: E0515 15:13:21.137796 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:21.140630 containerd[1549]: time="2025-05-15T15:13:21.140592694Z" level=info msg="connecting to shim f4e736525f881446a5987a4aedead0e1a1abf56b7809b2a3bd34a76ed8bcbc85" address="unix:///run/containerd/s/ecf854393674adc0ec17687ae5ae0c0f405df383c566396e63200a7018a88ff7" protocol=ttrpc version=3 May 15 15:13:21.141275 containerd[1549]: time="2025-05-15T15:13:21.141244728Z" level=info msg="CreateContainer within sandbox \"0b30890a417013c759b59380b1d9a38e1d8f4d53c2ac152686e21a83c7bc1b1d\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 15 15:13:21.150192 containerd[1549]: time="2025-05-15T15:13:21.150052211Z" level=info msg="Container 5d2927a7393b60d1d0de188df556f28638ef3c17fb143b3f6829a6f1cf654a44: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:21.163648 containerd[1549]: time="2025-05-15T15:13:21.163597094Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4334.0.0-a-073c24674a,Uid:b373e34f35214e049bae3b4082d9f52b,Namespace:kube-system,Attempt:0,} returns sandbox id \"900e02330d6a2d536094666bd2e5bece804162c265a360541832226b9718c8bd\"" May 15 15:13:21.165548 kubelet[2331]: E0515 15:13:21.165517 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:21.166217 systemd[1]: Started cri-containerd-f4e736525f881446a5987a4aedead0e1a1abf56b7809b2a3bd34a76ed8bcbc85.scope - libcontainer container f4e736525f881446a5987a4aedead0e1a1abf56b7809b2a3bd34a76ed8bcbc85. May 15 15:13:21.168301 containerd[1549]: time="2025-05-15T15:13:21.168087387Z" level=info msg="CreateContainer within sandbox \"900e02330d6a2d536094666bd2e5bece804162c265a360541832226b9718c8bd\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 15 15:13:21.169740 containerd[1549]: time="2025-05-15T15:13:21.169664248Z" level=info msg="CreateContainer within sandbox \"0b30890a417013c759b59380b1d9a38e1d8f4d53c2ac152686e21a83c7bc1b1d\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"5d2927a7393b60d1d0de188df556f28638ef3c17fb143b3f6829a6f1cf654a44\"" May 15 15:13:21.170446 containerd[1549]: time="2025-05-15T15:13:21.170120221Z" level=info msg="StartContainer for \"5d2927a7393b60d1d0de188df556f28638ef3c17fb143b3f6829a6f1cf654a44\"" May 15 15:13:21.173313 containerd[1549]: time="2025-05-15T15:13:21.172796613Z" level=info msg="connecting to shim 5d2927a7393b60d1d0de188df556f28638ef3c17fb143b3f6829a6f1cf654a44" address="unix:///run/containerd/s/8fe0f3a5d001f8ffa0c7bf145154e261427abb6c7459f3bfa6cad2aed2f7982e" protocol=ttrpc version=3 May 15 15:13:21.176967 containerd[1549]: time="2025-05-15T15:13:21.174994707Z" level=info msg="Container 6272d40bdc54c64c6bb00e7ae8975b2c9e125b567d20cb4a91d86df084f4bdbe: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:21.188421 containerd[1549]: time="2025-05-15T15:13:21.188373301Z" level=info msg="CreateContainer within sandbox \"900e02330d6a2d536094666bd2e5bece804162c265a360541832226b9718c8bd\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"6272d40bdc54c64c6bb00e7ae8975b2c9e125b567d20cb4a91d86df084f4bdbe\"" May 15 15:13:21.190226 containerd[1549]: time="2025-05-15T15:13:21.189000002Z" level=info msg="StartContainer for \"6272d40bdc54c64c6bb00e7ae8975b2c9e125b567d20cb4a91d86df084f4bdbe\"" May 15 15:13:21.193264 containerd[1549]: time="2025-05-15T15:13:21.193192751Z" level=info msg="connecting to shim 6272d40bdc54c64c6bb00e7ae8975b2c9e125b567d20cb4a91d86df084f4bdbe" address="unix:///run/containerd/s/b020c4387b1eb0b29cd54c0af9394ef9030b10ba013c7c6922ede5c97707e8dc" protocol=ttrpc version=3 May 15 15:13:21.213018 systemd[1]: Started cri-containerd-5d2927a7393b60d1d0de188df556f28638ef3c17fb143b3f6829a6f1cf654a44.scope - libcontainer container 5d2927a7393b60d1d0de188df556f28638ef3c17fb143b3f6829a6f1cf654a44. May 15 15:13:21.225407 systemd[1]: Started cri-containerd-6272d40bdc54c64c6bb00e7ae8975b2c9e125b567d20cb4a91d86df084f4bdbe.scope - libcontainer container 6272d40bdc54c64c6bb00e7ae8975b2c9e125b567d20cb4a91d86df084f4bdbe. May 15 15:13:21.270669 containerd[1549]: time="2025-05-15T15:13:21.270203338Z" level=info msg="StartContainer for \"f4e736525f881446a5987a4aedead0e1a1abf56b7809b2a3bd34a76ed8bcbc85\" returns successfully" May 15 15:13:21.309617 containerd[1549]: time="2025-05-15T15:13:21.309574510Z" level=info msg="StartContainer for \"6272d40bdc54c64c6bb00e7ae8975b2c9e125b567d20cb4a91d86df084f4bdbe\" returns successfully" May 15 15:13:21.344578 containerd[1549]: time="2025-05-15T15:13:21.344528473Z" level=info msg="StartContainer for \"5d2927a7393b60d1d0de188df556f28638ef3c17fb143b3f6829a6f1cf654a44\" returns successfully" May 15 15:13:21.353509 kubelet[2331]: W0515 15:13:21.353436 2331 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://24.144.90.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 24.144.90.91:6443: connect: connection refused May 15 15:13:21.353681 kubelet[2331]: E0515 15:13:21.353662 2331 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://24.144.90.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 24.144.90.91:6443: connect: connection refused" logger="UnhandledError" May 15 15:13:21.373326 kubelet[2331]: W0515 15:13:21.373079 2331 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://24.144.90.91:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4334.0.0-a-073c24674a&limit=500&resourceVersion=0": dial tcp 24.144.90.91:6443: connect: connection refused May 15 15:13:21.373326 kubelet[2331]: E0515 15:13:21.373175 2331 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://24.144.90.91:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4334.0.0-a-073c24674a&limit=500&resourceVersion=0\": dial tcp 24.144.90.91:6443: connect: connection refused" logger="UnhandledError" May 15 15:13:21.795993 kubelet[2331]: I0515 15:13:21.795398 2331 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-073c24674a" May 15 15:13:22.217371 kubelet[2331]: E0515 15:13:22.217341 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:22.217522 kubelet[2331]: E0515 15:13:22.217486 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:22.219261 kubelet[2331]: E0515 15:13:22.219231 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:22.219389 kubelet[2331]: E0515 15:13:22.219347 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:22.224535 kubelet[2331]: E0515 15:13:22.224504 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:22.224690 kubelet[2331]: E0515 15:13:22.224629 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:23.226545 kubelet[2331]: E0515 15:13:23.225893 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:23.226545 kubelet[2331]: E0515 15:13:23.226044 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:23.226545 kubelet[2331]: E0515 15:13:23.226261 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:23.226545 kubelet[2331]: E0515 15:13:23.226337 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:23.228366 kubelet[2331]: E0515 15:13:23.228234 2331 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:23.228457 kubelet[2331]: E0515 15:13:23.228445 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:23.561669 kubelet[2331]: E0515 15:13:23.561525 2331 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4334.0.0-a-073c24674a\" not found" node="ci-4334.0.0-a-073c24674a" May 15 15:13:23.696432 kubelet[2331]: I0515 15:13:23.695975 2331 kubelet_node_status.go:79] "Successfully registered node" node="ci-4334.0.0-a-073c24674a" May 15 15:13:23.758815 kubelet[2331]: I0515 15:13:23.758769 2331 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:23.765750 kubelet[2331]: E0515 15:13:23.765507 2331 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4334.0.0-a-073c24674a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:23.765750 kubelet[2331]: I0515 15:13:23.765549 2331 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:23.769962 kubelet[2331]: E0515 15:13:23.768784 2331 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:23.770414 kubelet[2331]: I0515 15:13:23.770178 2331 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:23.772418 kubelet[2331]: E0515 15:13:23.772387 2331 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4334.0.0-a-073c24674a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:24.132914 kubelet[2331]: I0515 15:13:24.132860 2331 apiserver.go:52] "Watching apiserver" May 15 15:13:24.158265 kubelet[2331]: I0515 15:13:24.158191 2331 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 15 15:13:24.229009 kubelet[2331]: I0515 15:13:24.227105 2331 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:24.229009 kubelet[2331]: I0515 15:13:24.227574 2331 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:24.229527 kubelet[2331]: E0515 15:13:24.229174 2331 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4334.0.0-a-073c24674a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:24.229527 kubelet[2331]: E0515 15:13:24.229368 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:24.230575 kubelet[2331]: E0515 15:13:24.230364 2331 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4334.0.0-a-073c24674a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:24.230575 kubelet[2331]: E0515 15:13:24.230521 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:25.229039 kubelet[2331]: I0515 15:13:25.228979 2331 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:25.238955 kubelet[2331]: W0515 15:13:25.238846 2331 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 15:13:25.239643 kubelet[2331]: E0515 15:13:25.239379 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:25.484605 systemd[1]: Reload requested from client PID 2600 ('systemctl') (unit session-7.scope)... May 15 15:13:25.484632 systemd[1]: Reloading... May 15 15:13:25.637972 zram_generator::config[2652]: No configuration found. May 15 15:13:25.652767 kubelet[2331]: I0515 15:13:25.652733 2331 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:25.660093 kubelet[2331]: W0515 15:13:25.660004 2331 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 15:13:25.660877 kubelet[2331]: E0515 15:13:25.660706 2331 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:25.790176 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 15 15:13:26.044650 systemd[1]: Reloading finished in 559 ms. May 15 15:13:26.089430 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 15 15:13:26.090305 kubelet[2331]: I0515 15:13:26.089805 2331 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 15 15:13:26.108559 systemd[1]: kubelet.service: Deactivated successfully. May 15 15:13:26.109289 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 15 15:13:26.109491 systemd[1]: kubelet.service: Consumed 1.058s CPU time, 120.8M memory peak. May 15 15:13:26.112787 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 15:13:26.300737 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 15 15:13:26.314475 (kubelet)[2694]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 15 15:13:26.385701 kubelet[2694]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 15 15:13:26.385701 kubelet[2694]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 15 15:13:26.385701 kubelet[2694]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 15 15:13:26.385701 kubelet[2694]: I0515 15:13:26.385483 2694 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 15 15:13:26.398312 kubelet[2694]: I0515 15:13:26.398238 2694 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" May 15 15:13:26.398312 kubelet[2694]: I0515 15:13:26.398277 2694 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 15 15:13:26.398718 kubelet[2694]: I0515 15:13:26.398605 2694 server.go:954] "Client rotation is on, will bootstrap in background" May 15 15:13:26.402218 kubelet[2694]: I0515 15:13:26.402185 2694 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 15 15:13:26.411991 kubelet[2694]: I0515 15:13:26.411852 2694 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 15 15:13:26.417626 kubelet[2694]: I0515 15:13:26.417578 2694 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 15 15:13:26.421732 kubelet[2694]: I0515 15:13:26.421694 2694 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 15 15:13:26.422218 kubelet[2694]: I0515 15:13:26.422107 2694 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 15 15:13:26.422390 kubelet[2694]: I0515 15:13:26.422143 2694 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4334.0.0-a-073c24674a","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 15 15:13:26.422390 kubelet[2694]: I0515 15:13:26.422355 2694 topology_manager.go:138] "Creating topology manager with none policy" May 15 15:13:26.422390 kubelet[2694]: I0515 15:13:26.422364 2694 container_manager_linux.go:304] "Creating device plugin manager" May 15 15:13:26.422739 kubelet[2694]: I0515 15:13:26.422409 2694 state_mem.go:36] "Initialized new in-memory state store" May 15 15:13:26.422739 kubelet[2694]: I0515 15:13:26.422601 2694 kubelet.go:446] "Attempting to sync node with API server" May 15 15:13:26.422739 kubelet[2694]: I0515 15:13:26.422620 2694 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 15 15:13:26.423348 kubelet[2694]: I0515 15:13:26.423218 2694 kubelet.go:352] "Adding apiserver pod source" May 15 15:13:26.423348 kubelet[2694]: I0515 15:13:26.423254 2694 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 15 15:13:26.426759 kubelet[2694]: I0515 15:13:26.426696 2694 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 15 15:13:26.429129 kubelet[2694]: I0515 15:13:26.429081 2694 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 15 15:13:26.434037 kubelet[2694]: I0515 15:13:26.433964 2694 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 15 15:13:26.434337 kubelet[2694]: I0515 15:13:26.434150 2694 server.go:1287] "Started kubelet" May 15 15:13:26.455397 kubelet[2694]: I0515 15:13:26.455269 2694 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 15 15:13:26.456870 kubelet[2694]: I0515 15:13:26.456183 2694 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 15 15:13:26.457359 kubelet[2694]: I0515 15:13:26.457333 2694 server.go:490] "Adding debug handlers to kubelet server" May 15 15:13:26.460736 kubelet[2694]: I0515 15:13:26.460667 2694 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 15 15:13:26.460875 kubelet[2694]: I0515 15:13:26.460860 2694 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 15 15:13:26.463054 kubelet[2694]: I0515 15:13:26.462776 2694 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 15 15:13:26.467398 kubelet[2694]: I0515 15:13:26.467218 2694 volume_manager.go:297] "Starting Kubelet Volume Manager" May 15 15:13:26.468472 kubelet[2694]: I0515 15:13:26.468284 2694 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 15 15:13:26.470302 kubelet[2694]: I0515 15:13:26.469146 2694 reconciler.go:26] "Reconciler: start to sync state" May 15 15:13:26.474692 kubelet[2694]: I0515 15:13:26.474326 2694 factory.go:221] Registration of the systemd container factory successfully May 15 15:13:26.476201 kubelet[2694]: I0515 15:13:26.476147 2694 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 15 15:13:26.478288 kubelet[2694]: I0515 15:13:26.477888 2694 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 15 15:13:26.482052 kubelet[2694]: I0515 15:13:26.481502 2694 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 15 15:13:26.482052 kubelet[2694]: I0515 15:13:26.481532 2694 status_manager.go:227] "Starting to sync pod status with apiserver" May 15 15:13:26.482052 kubelet[2694]: I0515 15:13:26.481554 2694 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 15 15:13:26.482052 kubelet[2694]: I0515 15:13:26.481562 2694 kubelet.go:2388] "Starting kubelet main sync loop" May 15 15:13:26.482052 kubelet[2694]: E0515 15:13:26.481612 2694 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 15 15:13:26.484501 kubelet[2694]: E0515 15:13:26.484470 2694 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 15 15:13:26.489699 kubelet[2694]: I0515 15:13:26.489664 2694 factory.go:221] Registration of the containerd container factory successfully May 15 15:13:26.579210 kubelet[2694]: I0515 15:13:26.579073 2694 cpu_manager.go:221] "Starting CPU manager" policy="none" May 15 15:13:26.579421 kubelet[2694]: I0515 15:13:26.579398 2694 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 15 15:13:26.579893 kubelet[2694]: I0515 15:13:26.579511 2694 state_mem.go:36] "Initialized new in-memory state store" May 15 15:13:26.579893 kubelet[2694]: I0515 15:13:26.579750 2694 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 15 15:13:26.579893 kubelet[2694]: I0515 15:13:26.579766 2694 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 15 15:13:26.579893 kubelet[2694]: I0515 15:13:26.579796 2694 policy_none.go:49] "None policy: Start" May 15 15:13:26.579893 kubelet[2694]: I0515 15:13:26.579809 2694 memory_manager.go:186] "Starting memorymanager" policy="None" May 15 15:13:26.579893 kubelet[2694]: I0515 15:13:26.579823 2694 state_mem.go:35] "Initializing new in-memory state store" May 15 15:13:26.581113 kubelet[2694]: I0515 15:13:26.581023 2694 state_mem.go:75] "Updated machine memory state" May 15 15:13:26.581735 kubelet[2694]: E0515 15:13:26.581707 2694 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 15 15:13:26.599728 kubelet[2694]: I0515 15:13:26.595221 2694 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 15 15:13:26.599728 kubelet[2694]: I0515 15:13:26.595474 2694 eviction_manager.go:189] "Eviction manager: starting control loop" May 15 15:13:26.599728 kubelet[2694]: I0515 15:13:26.595491 2694 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 15 15:13:26.599728 kubelet[2694]: I0515 15:13:26.595910 2694 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 15 15:13:26.603473 kubelet[2694]: E0515 15:13:26.603443 2694 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 15 15:13:26.710694 kubelet[2694]: I0515 15:13:26.710663 2694 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-073c24674a" May 15 15:13:26.722319 kubelet[2694]: I0515 15:13:26.722287 2694 kubelet_node_status.go:125] "Node was previously registered" node="ci-4334.0.0-a-073c24674a" May 15 15:13:26.722601 kubelet[2694]: I0515 15:13:26.722581 2694 kubelet_node_status.go:79] "Successfully registered node" node="ci-4334.0.0-a-073c24674a" May 15 15:13:26.783002 kubelet[2694]: I0515 15:13:26.782704 2694 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:26.786457 kubelet[2694]: I0515 15:13:26.785893 2694 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:26.786457 kubelet[2694]: I0515 15:13:26.786031 2694 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:26.797659 kubelet[2694]: W0515 15:13:26.797448 2694 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 15:13:26.797659 kubelet[2694]: E0515 15:13:26.797522 2694 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4334.0.0-a-073c24674a\" already exists" pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:26.797807 kubelet[2694]: W0515 15:13:26.797690 2694 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 15:13:26.800569 kubelet[2694]: W0515 15:13:26.800505 2694 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 15:13:26.800569 kubelet[2694]: E0515 15:13:26.800565 2694 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4334.0.0-a-073c24674a\" already exists" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872565 kubelet[2694]: I0515 15:13:26.872128 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0e261e850a4cc2912c11fa835e595fef-kubeconfig\") pod \"kube-scheduler-ci-4334.0.0-a-073c24674a\" (UID: \"0e261e850a4cc2912c11fa835e595fef\") " pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872565 kubelet[2694]: I0515 15:13:26.872213 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b49fd5be0a8bc73cf8e412ce96b1971d-k8s-certs\") pod \"kube-apiserver-ci-4334.0.0-a-073c24674a\" (UID: \"b49fd5be0a8bc73cf8e412ce96b1971d\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872565 kubelet[2694]: I0515 15:13:26.872236 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b49fd5be0a8bc73cf8e412ce96b1971d-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4334.0.0-a-073c24674a\" (UID: \"b49fd5be0a8bc73cf8e412ce96b1971d\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872565 kubelet[2694]: I0515 15:13:26.872255 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-kubeconfig\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872565 kubelet[2694]: I0515 15:13:26.872311 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872860 kubelet[2694]: I0515 15:13:26.872325 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b49fd5be0a8bc73cf8e412ce96b1971d-ca-certs\") pod \"kube-apiserver-ci-4334.0.0-a-073c24674a\" (UID: \"b49fd5be0a8bc73cf8e412ce96b1971d\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872860 kubelet[2694]: I0515 15:13:26.872365 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-ca-certs\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872860 kubelet[2694]: I0515 15:13:26.872384 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-flexvolume-dir\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:26.872860 kubelet[2694]: I0515 15:13:26.872421 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b373e34f35214e049bae3b4082d9f52b-k8s-certs\") pod \"kube-controller-manager-ci-4334.0.0-a-073c24674a\" (UID: \"b373e34f35214e049bae3b4082d9f52b\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" May 15 15:13:27.097949 kubelet[2694]: E0515 15:13:27.097876 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:27.098743 kubelet[2694]: E0515 15:13:27.098424 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:27.101690 kubelet[2694]: E0515 15:13:27.101361 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:27.426995 kubelet[2694]: I0515 15:13:27.426405 2694 apiserver.go:52] "Watching apiserver" May 15 15:13:27.469417 kubelet[2694]: I0515 15:13:27.469355 2694 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 15 15:13:27.538352 kubelet[2694]: E0515 15:13:27.538232 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:27.539364 kubelet[2694]: E0515 15:13:27.539319 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:27.539844 kubelet[2694]: I0515 15:13:27.539790 2694 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:27.583003 kubelet[2694]: W0515 15:13:27.582974 2694 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 15:13:27.585363 kubelet[2694]: E0515 15:13:27.585046 2694 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4334.0.0-a-073c24674a\" already exists" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" May 15 15:13:27.585363 kubelet[2694]: E0515 15:13:27.585243 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:27.594239 kubelet[2694]: I0515 15:13:27.594167 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4334.0.0-a-073c24674a" podStartSLOduration=2.594114771 podStartE2EDuration="2.594114771s" podCreationTimestamp="2025-05-15 15:13:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:13:27.550796418 +0000 UTC m=+1.231669445" watchObservedRunningTime="2025-05-15 15:13:27.594114771 +0000 UTC m=+1.274987796" May 15 15:13:27.648260 kubelet[2694]: I0515 15:13:27.647847 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-073c24674a" podStartSLOduration=1.647824982 podStartE2EDuration="1.647824982s" podCreationTimestamp="2025-05-15 15:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:13:27.594810859 +0000 UTC m=+1.275683866" watchObservedRunningTime="2025-05-15 15:13:27.647824982 +0000 UTC m=+1.328698008" May 15 15:13:27.679507 kubelet[2694]: I0515 15:13:27.679011 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4334.0.0-a-073c24674a" podStartSLOduration=2.6789920130000002 podStartE2EDuration="2.678992013s" podCreationTimestamp="2025-05-15 15:13:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:13:27.64897109 +0000 UTC m=+1.329844116" watchObservedRunningTime="2025-05-15 15:13:27.678992013 +0000 UTC m=+1.359865040" May 15 15:13:28.542510 kubelet[2694]: E0515 15:13:28.542452 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:28.544070 kubelet[2694]: E0515 15:13:28.544044 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:28.784213 systemd[1]: Started sshd@7-24.144.90.91:22-128.1.132.137:55584.service - OpenSSH per-connection server daemon (128.1.132.137:55584). May 15 15:13:29.548222 kubelet[2694]: E0515 15:13:29.548184 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:29.671488 sshd[2740]: Invalid user eloa from 128.1.132.137 port 55584 May 15 15:13:29.837112 sshd[2740]: Received disconnect from 128.1.132.137 port 55584:11: Bye Bye [preauth] May 15 15:13:29.837112 sshd[2740]: Disconnected from invalid user eloa 128.1.132.137 port 55584 [preauth] May 15 15:13:29.839743 systemd[1]: sshd@7-24.144.90.91:22-128.1.132.137:55584.service: Deactivated successfully. May 15 15:13:30.264815 kubelet[2694]: I0515 15:13:30.264697 2694 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 15 15:13:30.265364 containerd[1549]: time="2025-05-15T15:13:30.265317582Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 15 15:13:30.265846 kubelet[2694]: I0515 15:13:30.265610 2694 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 15 15:13:30.549855 kubelet[2694]: E0515 15:13:30.549552 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:31.244881 systemd[1]: Created slice kubepods-besteffort-pod70fb20dd_b97e_40f3_992d_acae9d925172.slice - libcontainer container kubepods-besteffort-pod70fb20dd_b97e_40f3_992d_acae9d925172.slice. May 15 15:13:31.305949 kubelet[2694]: I0515 15:13:31.305874 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/70fb20dd-b97e-40f3-992d-acae9d925172-xtables-lock\") pod \"kube-proxy-lq2cg\" (UID: \"70fb20dd-b97e-40f3-992d-acae9d925172\") " pod="kube-system/kube-proxy-lq2cg" May 15 15:13:31.305949 kubelet[2694]: I0515 15:13:31.305915 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/70fb20dd-b97e-40f3-992d-acae9d925172-kube-proxy\") pod \"kube-proxy-lq2cg\" (UID: \"70fb20dd-b97e-40f3-992d-acae9d925172\") " pod="kube-system/kube-proxy-lq2cg" May 15 15:13:31.305949 kubelet[2694]: I0515 15:13:31.305952 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q47nt\" (UniqueName: \"kubernetes.io/projected/70fb20dd-b97e-40f3-992d-acae9d925172-kube-api-access-q47nt\") pod \"kube-proxy-lq2cg\" (UID: \"70fb20dd-b97e-40f3-992d-acae9d925172\") " pod="kube-system/kube-proxy-lq2cg" May 15 15:13:31.306171 kubelet[2694]: I0515 15:13:31.305973 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/70fb20dd-b97e-40f3-992d-acae9d925172-lib-modules\") pod \"kube-proxy-lq2cg\" (UID: \"70fb20dd-b97e-40f3-992d-acae9d925172\") " pod="kube-system/kube-proxy-lq2cg" May 15 15:13:31.402338 systemd[1]: Created slice kubepods-besteffort-pode6428a27_8402_49ec_8695_66990e614a22.slice - libcontainer container kubepods-besteffort-pode6428a27_8402_49ec_8695_66990e614a22.slice. May 15 15:13:31.412886 kubelet[2694]: I0515 15:13:31.407148 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/e6428a27-8402-49ec-8695-66990e614a22-var-lib-calico\") pod \"tigera-operator-789496d6f5-tcvwc\" (UID: \"e6428a27-8402-49ec-8695-66990e614a22\") " pod="tigera-operator/tigera-operator-789496d6f5-tcvwc" May 15 15:13:31.413220 kubelet[2694]: I0515 15:13:31.413148 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9d5x\" (UniqueName: \"kubernetes.io/projected/e6428a27-8402-49ec-8695-66990e614a22-kube-api-access-k9d5x\") pod \"tigera-operator-789496d6f5-tcvwc\" (UID: \"e6428a27-8402-49ec-8695-66990e614a22\") " pod="tigera-operator/tigera-operator-789496d6f5-tcvwc" May 15 15:13:31.551747 kubelet[2694]: E0515 15:13:31.551498 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:31.553295 containerd[1549]: time="2025-05-15T15:13:31.553205941Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-lq2cg,Uid:70fb20dd-b97e-40f3-992d-acae9d925172,Namespace:kube-system,Attempt:0,}" May 15 15:13:31.571841 containerd[1549]: time="2025-05-15T15:13:31.571519080Z" level=info msg="connecting to shim 753c4e21403b24f5c922f4f76bce3b28b8596bf257052b22dd4b5f7a18adf9ed" address="unix:///run/containerd/s/04f052da9393a14fcb2072b8beddd1947450837fa4c90981ca052971c71dac03" namespace=k8s.io protocol=ttrpc version=3 May 15 15:13:31.606151 systemd[1]: Started cri-containerd-753c4e21403b24f5c922f4f76bce3b28b8596bf257052b22dd4b5f7a18adf9ed.scope - libcontainer container 753c4e21403b24f5c922f4f76bce3b28b8596bf257052b22dd4b5f7a18adf9ed. May 15 15:13:31.639158 containerd[1549]: time="2025-05-15T15:13:31.639107153Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-lq2cg,Uid:70fb20dd-b97e-40f3-992d-acae9d925172,Namespace:kube-system,Attempt:0,} returns sandbox id \"753c4e21403b24f5c922f4f76bce3b28b8596bf257052b22dd4b5f7a18adf9ed\"" May 15 15:13:31.640161 kubelet[2694]: E0515 15:13:31.640115 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:31.645052 containerd[1549]: time="2025-05-15T15:13:31.644278053Z" level=info msg="CreateContainer within sandbox \"753c4e21403b24f5c922f4f76bce3b28b8596bf257052b22dd4b5f7a18adf9ed\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 15 15:13:31.653113 containerd[1549]: time="2025-05-15T15:13:31.653073015Z" level=info msg="Container 2a8f65c23f264a97159571d790fc85956977ee993491d6136189ed6be52fa6a2: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:31.659820 containerd[1549]: time="2025-05-15T15:13:31.659748307Z" level=info msg="CreateContainer within sandbox \"753c4e21403b24f5c922f4f76bce3b28b8596bf257052b22dd4b5f7a18adf9ed\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"2a8f65c23f264a97159571d790fc85956977ee993491d6136189ed6be52fa6a2\"" May 15 15:13:31.661306 containerd[1549]: time="2025-05-15T15:13:31.661218427Z" level=info msg="StartContainer for \"2a8f65c23f264a97159571d790fc85956977ee993491d6136189ed6be52fa6a2\"" May 15 15:13:31.663295 containerd[1549]: time="2025-05-15T15:13:31.663263532Z" level=info msg="connecting to shim 2a8f65c23f264a97159571d790fc85956977ee993491d6136189ed6be52fa6a2" address="unix:///run/containerd/s/04f052da9393a14fcb2072b8beddd1947450837fa4c90981ca052971c71dac03" protocol=ttrpc version=3 May 15 15:13:31.698270 systemd[1]: Started cri-containerd-2a8f65c23f264a97159571d790fc85956977ee993491d6136189ed6be52fa6a2.scope - libcontainer container 2a8f65c23f264a97159571d790fc85956977ee993491d6136189ed6be52fa6a2. May 15 15:13:31.716054 containerd[1549]: time="2025-05-15T15:13:31.716006840Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-789496d6f5-tcvwc,Uid:e6428a27-8402-49ec-8695-66990e614a22,Namespace:tigera-operator,Attempt:0,}" May 15 15:13:31.748537 containerd[1549]: time="2025-05-15T15:13:31.748473862Z" level=info msg="connecting to shim f4ac77573811414037f0a7702ece63b235a7745bb5fc45c28e3fba39c5786831" address="unix:///run/containerd/s/ee3fdb26ecdbed7472651defd92828c3c0c008ac07470914bad30cc39b3ee8c8" namespace=k8s.io protocol=ttrpc version=3 May 15 15:13:31.762272 containerd[1549]: time="2025-05-15T15:13:31.762206477Z" level=info msg="StartContainer for \"2a8f65c23f264a97159571d790fc85956977ee993491d6136189ed6be52fa6a2\" returns successfully" May 15 15:13:31.795548 systemd[1]: Started cri-containerd-f4ac77573811414037f0a7702ece63b235a7745bb5fc45c28e3fba39c5786831.scope - libcontainer container f4ac77573811414037f0a7702ece63b235a7745bb5fc45c28e3fba39c5786831. May 15 15:13:31.875911 containerd[1549]: time="2025-05-15T15:13:31.875395478Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-789496d6f5-tcvwc,Uid:e6428a27-8402-49ec-8695-66990e614a22,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"f4ac77573811414037f0a7702ece63b235a7745bb5fc45c28e3fba39c5786831\"" May 15 15:13:31.881575 containerd[1549]: time="2025-05-15T15:13:31.881403815Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" May 15 15:13:31.888297 systemd-resolved[1408]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. May 15 15:13:31.950510 sudo[1782]: pam_unix(sudo:session): session closed for user root May 15 15:13:31.953711 sshd[1781]: Connection closed by 139.178.68.195 port 57670 May 15 15:13:31.954625 sshd-session[1779]: pam_unix(sshd:session): session closed for user core May 15 15:13:31.961457 systemd-logind[1527]: Session 7 logged out. Waiting for processes to exit. May 15 15:13:31.962188 systemd[1]: sshd@6-24.144.90.91:22-139.178.68.195:57670.service: Deactivated successfully. May 15 15:13:31.965793 systemd[1]: session-7.scope: Deactivated successfully. May 15 15:13:31.966507 systemd[1]: session-7.scope: Consumed 5.180s CPU time, 159.9M memory peak. May 15 15:13:31.969947 systemd-logind[1527]: Removed session 7. May 15 15:13:32.556376 kubelet[2694]: E0515 15:13:32.555349 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:32.566378 kubelet[2694]: I0515 15:13:32.566179 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-lq2cg" podStartSLOduration=1.5661537490000002 podStartE2EDuration="1.566153749s" podCreationTimestamp="2025-05-15 15:13:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:13:32.565639081 +0000 UTC m=+6.246512109" watchObservedRunningTime="2025-05-15 15:13:32.566153749 +0000 UTC m=+6.247026758" May 15 15:13:34.295066 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount566101366.mount: Deactivated successfully. May 15 15:13:34.302796 kubelet[2694]: E0515 15:13:34.302479 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:34.565385 kubelet[2694]: E0515 15:13:34.565036 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:34.819408 containerd[1549]: time="2025-05-15T15:13:34.818410214Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:34.819408 containerd[1549]: time="2025-05-15T15:13:34.819001820Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.7: active requests=0, bytes read=22002662" May 15 15:13:34.819408 containerd[1549]: time="2025-05-15T15:13:34.819069481Z" level=info msg="ImageCreate event name:\"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:34.820984 containerd[1549]: time="2025-05-15T15:13:34.820950033Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:34.821920 containerd[1549]: time="2025-05-15T15:13:34.821886634Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.7\" with image id \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\", repo tag \"quay.io/tigera/operator:v1.36.7\", repo digest \"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\", size \"21998657\" in 2.940069128s" May 15 15:13:34.822094 containerd[1549]: time="2025-05-15T15:13:34.822076191Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\"" May 15 15:13:34.824759 containerd[1549]: time="2025-05-15T15:13:34.824436111Z" level=info msg="CreateContainer within sandbox \"f4ac77573811414037f0a7702ece63b235a7745bb5fc45c28e3fba39c5786831\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" May 15 15:13:34.832912 containerd[1549]: time="2025-05-15T15:13:34.832297843Z" level=info msg="Container b5f5620fe3b7591f50b656584b3a3d92e326792c7e1cc4cdf0ed1bbb09e73cfd: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:34.840639 containerd[1549]: time="2025-05-15T15:13:34.840597701Z" level=info msg="CreateContainer within sandbox \"f4ac77573811414037f0a7702ece63b235a7745bb5fc45c28e3fba39c5786831\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"b5f5620fe3b7591f50b656584b3a3d92e326792c7e1cc4cdf0ed1bbb09e73cfd\"" May 15 15:13:34.843040 containerd[1549]: time="2025-05-15T15:13:34.842995736Z" level=info msg="StartContainer for \"b5f5620fe3b7591f50b656584b3a3d92e326792c7e1cc4cdf0ed1bbb09e73cfd\"" May 15 15:13:34.844427 containerd[1549]: time="2025-05-15T15:13:34.844367497Z" level=info msg="connecting to shim b5f5620fe3b7591f50b656584b3a3d92e326792c7e1cc4cdf0ed1bbb09e73cfd" address="unix:///run/containerd/s/ee3fdb26ecdbed7472651defd92828c3c0c008ac07470914bad30cc39b3ee8c8" protocol=ttrpc version=3 May 15 15:13:34.873224 systemd[1]: Started cri-containerd-b5f5620fe3b7591f50b656584b3a3d92e326792c7e1cc4cdf0ed1bbb09e73cfd.scope - libcontainer container b5f5620fe3b7591f50b656584b3a3d92e326792c7e1cc4cdf0ed1bbb09e73cfd. May 15 15:13:34.906896 containerd[1549]: time="2025-05-15T15:13:34.906806012Z" level=info msg="StartContainer for \"b5f5620fe3b7591f50b656584b3a3d92e326792c7e1cc4cdf0ed1bbb09e73cfd\" returns successfully" May 15 15:13:35.497988 kubelet[2694]: E0515 15:13:35.497570 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:35.569049 kubelet[2694]: E0515 15:13:35.568887 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:35.591961 kubelet[2694]: I0515 15:13:35.591891 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-789496d6f5-tcvwc" podStartSLOduration=1.6493657430000002 podStartE2EDuration="4.591862822s" podCreationTimestamp="2025-05-15 15:13:31 +0000 UTC" firstStartedPulling="2025-05-15 15:13:31.880478947 +0000 UTC m=+5.561351952" lastFinishedPulling="2025-05-15 15:13:34.822976026 +0000 UTC m=+8.503849031" observedRunningTime="2025-05-15 15:13:35.590923137 +0000 UTC m=+9.271796163" watchObservedRunningTime="2025-05-15 15:13:35.591862822 +0000 UTC m=+9.272735847" May 15 15:13:38.179686 systemd[1]: Created slice kubepods-besteffort-pod12abe38c_3744_42fb_bd9c_1466a933b3e0.slice - libcontainer container kubepods-besteffort-pod12abe38c_3744_42fb_bd9c_1466a933b3e0.slice. May 15 15:13:38.260006 kubelet[2694]: I0515 15:13:38.259178 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12abe38c-3744-42fb-bd9c-1466a933b3e0-tigera-ca-bundle\") pod \"calico-typha-596f4965b7-vjc2l\" (UID: \"12abe38c-3744-42fb-bd9c-1466a933b3e0\") " pod="calico-system/calico-typha-596f4965b7-vjc2l" May 15 15:13:38.260006 kubelet[2694]: I0515 15:13:38.259258 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/12abe38c-3744-42fb-bd9c-1466a933b3e0-typha-certs\") pod \"calico-typha-596f4965b7-vjc2l\" (UID: \"12abe38c-3744-42fb-bd9c-1466a933b3e0\") " pod="calico-system/calico-typha-596f4965b7-vjc2l" May 15 15:13:38.260006 kubelet[2694]: I0515 15:13:38.259292 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkwfq\" (UniqueName: \"kubernetes.io/projected/12abe38c-3744-42fb-bd9c-1466a933b3e0-kube-api-access-kkwfq\") pod \"calico-typha-596f4965b7-vjc2l\" (UID: \"12abe38c-3744-42fb-bd9c-1466a933b3e0\") " pod="calico-system/calico-typha-596f4965b7-vjc2l" May 15 15:13:38.372205 systemd[1]: Created slice kubepods-besteffort-pod06424021_71ea_4cef_bbec_94f6e5d79ae3.slice - libcontainer container kubepods-besteffort-pod06424021_71ea_4cef_bbec_94f6e5d79ae3.slice. May 15 15:13:38.462398 kubelet[2694]: I0515 15:13:38.462272 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghrtz\" (UniqueName: \"kubernetes.io/projected/06424021-71ea-4cef-bbec-94f6e5d79ae3-kube-api-access-ghrtz\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.462851 kubelet[2694]: I0515 15:13:38.462630 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-var-run-calico\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.462996 kubelet[2694]: I0515 15:13:38.462981 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-policysync\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463126 kubelet[2694]: I0515 15:13:38.463074 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/06424021-71ea-4cef-bbec-94f6e5d79ae3-node-certs\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463126 kubelet[2694]: I0515 15:13:38.463092 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-net-dir\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463126 kubelet[2694]: I0515 15:13:38.463107 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-flexvol-driver-host\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463300 kubelet[2694]: I0515 15:13:38.463239 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-xtables-lock\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463300 kubelet[2694]: I0515 15:13:38.463264 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-var-lib-calico\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463300 kubelet[2694]: I0515 15:13:38.463276 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-bin-dir\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463953 kubelet[2694]: I0515 15:13:38.463405 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06424021-71ea-4cef-bbec-94f6e5d79ae3-tigera-ca-bundle\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463953 kubelet[2694]: I0515 15:13:38.463429 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-lib-modules\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.463953 kubelet[2694]: I0515 15:13:38.463445 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-log-dir\") pod \"calico-node-qjmm6\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " pod="calico-system/calico-node-qjmm6" May 15 15:13:38.474295 kubelet[2694]: E0515 15:13:38.473570 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5mw49" podUID="ea7342ed-f869-4f54-b67c-23b40f5f8961" May 15 15:13:38.485386 kubelet[2694]: E0515 15:13:38.485356 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:38.486447 containerd[1549]: time="2025-05-15T15:13:38.486392820Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-596f4965b7-vjc2l,Uid:12abe38c-3744-42fb-bd9c-1466a933b3e0,Namespace:calico-system,Attempt:0,}" May 15 15:13:38.519206 containerd[1549]: time="2025-05-15T15:13:38.519005388Z" level=info msg="connecting to shim 4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58" address="unix:///run/containerd/s/757fbf1bcad6294ce35343044f403e61e29c683483b74643c9d8783cc603dd74" namespace=k8s.io protocol=ttrpc version=3 May 15 15:13:38.560204 systemd[1]: Started cri-containerd-4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58.scope - libcontainer container 4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58. May 15 15:13:38.565172 kubelet[2694]: I0515 15:13:38.565105 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2492b\" (UniqueName: \"kubernetes.io/projected/ea7342ed-f869-4f54-b67c-23b40f5f8961-kube-api-access-2492b\") pod \"csi-node-driver-5mw49\" (UID: \"ea7342ed-f869-4f54-b67c-23b40f5f8961\") " pod="calico-system/csi-node-driver-5mw49" May 15 15:13:38.565172 kubelet[2694]: I0515 15:13:38.565151 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ea7342ed-f869-4f54-b67c-23b40f5f8961-registration-dir\") pod \"csi-node-driver-5mw49\" (UID: \"ea7342ed-f869-4f54-b67c-23b40f5f8961\") " pod="calico-system/csi-node-driver-5mw49" May 15 15:13:38.566678 kubelet[2694]: I0515 15:13:38.565907 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/ea7342ed-f869-4f54-b67c-23b40f5f8961-varrun\") pod \"csi-node-driver-5mw49\" (UID: \"ea7342ed-f869-4f54-b67c-23b40f5f8961\") " pod="calico-system/csi-node-driver-5mw49" May 15 15:13:38.566678 kubelet[2694]: I0515 15:13:38.565962 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea7342ed-f869-4f54-b67c-23b40f5f8961-kubelet-dir\") pod \"csi-node-driver-5mw49\" (UID: \"ea7342ed-f869-4f54-b67c-23b40f5f8961\") " pod="calico-system/csi-node-driver-5mw49" May 15 15:13:38.566678 kubelet[2694]: I0515 15:13:38.566081 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ea7342ed-f869-4f54-b67c-23b40f5f8961-socket-dir\") pod \"csi-node-driver-5mw49\" (UID: \"ea7342ed-f869-4f54-b67c-23b40f5f8961\") " pod="calico-system/csi-node-driver-5mw49" May 15 15:13:38.569206 kubelet[2694]: E0515 15:13:38.569174 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.569414 kubelet[2694]: W0515 15:13:38.569390 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.569567 kubelet[2694]: E0515 15:13:38.569548 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.571156 kubelet[2694]: E0515 15:13:38.571134 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.571293 kubelet[2694]: W0515 15:13:38.571274 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.571512 kubelet[2694]: E0515 15:13:38.571493 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.572241 kubelet[2694]: E0515 15:13:38.572206 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.572468 kubelet[2694]: W0515 15:13:38.572224 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.572626 kubelet[2694]: E0515 15:13:38.572527 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.573351 kubelet[2694]: E0515 15:13:38.573179 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.573351 kubelet[2694]: W0515 15:13:38.573305 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.573351 kubelet[2694]: E0515 15:13:38.573323 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.581098 kubelet[2694]: E0515 15:13:38.579492 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.581098 kubelet[2694]: W0515 15:13:38.580967 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.581098 kubelet[2694]: E0515 15:13:38.581011 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.584403 kubelet[2694]: E0515 15:13:38.584286 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.584403 kubelet[2694]: W0515 15:13:38.584312 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.584403 kubelet[2694]: E0515 15:13:38.584339 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.602976 kubelet[2694]: E0515 15:13:38.602830 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.602976 kubelet[2694]: W0515 15:13:38.602861 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.602976 kubelet[2694]: E0515 15:13:38.602890 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.667543 kubelet[2694]: E0515 15:13:38.667326 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.667543 kubelet[2694]: W0515 15:13:38.667355 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.667543 kubelet[2694]: E0515 15:13:38.667381 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.668311 kubelet[2694]: E0515 15:13:38.668292 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.668609 kubelet[2694]: W0515 15:13:38.668432 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.668609 kubelet[2694]: E0515 15:13:38.668458 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.668821 kubelet[2694]: E0515 15:13:38.668800 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.669162 kubelet[2694]: W0515 15:13:38.668956 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.669162 kubelet[2694]: E0515 15:13:38.669013 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.670098 kubelet[2694]: E0515 15:13:38.670067 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.670098 kubelet[2694]: W0515 15:13:38.670087 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.670589 kubelet[2694]: E0515 15:13:38.670116 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.670589 kubelet[2694]: E0515 15:13:38.670442 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.670589 kubelet[2694]: W0515 15:13:38.670455 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.670589 kubelet[2694]: E0515 15:13:38.670497 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.671191 kubelet[2694]: E0515 15:13:38.670647 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.671191 kubelet[2694]: W0515 15:13:38.670658 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.671191 kubelet[2694]: E0515 15:13:38.670816 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.671191 kubelet[2694]: E0515 15:13:38.670847 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.671191 kubelet[2694]: W0515 15:13:38.670857 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.671191 kubelet[2694]: E0515 15:13:38.670882 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.671838 kubelet[2694]: E0515 15:13:38.671821 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.672093 kubelet[2694]: W0515 15:13:38.671843 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.672093 kubelet[2694]: E0515 15:13:38.671958 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.672093 kubelet[2694]: E0515 15:13:38.672068 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.672093 kubelet[2694]: W0515 15:13:38.672078 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.673187 kubelet[2694]: E0515 15:13:38.673076 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.673187 kubelet[2694]: W0515 15:13:38.673089 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.673382 kubelet[2694]: E0515 15:13:38.673324 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.673382 kubelet[2694]: E0515 15:13:38.673361 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.674094 kubelet[2694]: E0515 15:13:38.674068 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.674094 kubelet[2694]: W0515 15:13:38.674083 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.674371 kubelet[2694]: E0515 15:13:38.674151 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.674371 kubelet[2694]: E0515 15:13:38.674280 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.674371 kubelet[2694]: W0515 15:13:38.674291 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.674587 kubelet[2694]: E0515 15:13:38.674480 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.674587 kubelet[2694]: W0515 15:13:38.674491 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.674841 kubelet[2694]: E0515 15:13:38.674695 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.674841 kubelet[2694]: E0515 15:13:38.674722 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.674841 kubelet[2694]: E0515 15:13:38.674725 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.674841 kubelet[2694]: W0515 15:13:38.674732 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.674841 kubelet[2694]: E0515 15:13:38.674759 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.675332 kubelet[2694]: E0515 15:13:38.675130 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.675332 kubelet[2694]: W0515 15:13:38.675142 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.676023 kubelet[2694]: E0515 15:13:38.675835 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.676335 kubelet[2694]: E0515 15:13:38.675924 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.677770 kubelet[2694]: W0515 15:13:38.677740 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.677901 kubelet[2694]: E0515 15:13:38.677887 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.678240 kubelet[2694]: E0515 15:13:38.678227 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.678331 kubelet[2694]: W0515 15:13:38.678320 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.678418 kubelet[2694]: E0515 15:13:38.678403 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.678659 kubelet[2694]: E0515 15:13:38.678637 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.678659 kubelet[2694]: W0515 15:13:38.678653 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.678809 kubelet[2694]: E0515 15:13:38.678675 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.679007 kubelet[2694]: E0515 15:13:38.678987 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.679007 kubelet[2694]: W0515 15:13:38.679005 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.679222 kubelet[2694]: E0515 15:13:38.679105 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.679977 kubelet[2694]: E0515 15:13:38.679957 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.679977 kubelet[2694]: W0515 15:13:38.679975 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.680383 kubelet[2694]: E0515 15:13:38.680143 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.680383 kubelet[2694]: E0515 15:13:38.680176 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.680383 kubelet[2694]: W0515 15:13:38.680187 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.680943 kubelet[2694]: E0515 15:13:38.680908 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.681023 kubelet[2694]: E0515 15:13:38.680949 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.681023 kubelet[2694]: W0515 15:13:38.680961 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.681023 kubelet[2694]: E0515 15:13:38.680975 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.681543 kubelet[2694]: E0515 15:13:38.681249 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.681543 kubelet[2694]: W0515 15:13:38.681266 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.681543 kubelet[2694]: E0515 15:13:38.681292 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.682828 kubelet[2694]: E0515 15:13:38.682714 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.682828 kubelet[2694]: W0515 15:13:38.682733 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.682828 kubelet[2694]: E0515 15:13:38.682747 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.684191 kubelet[2694]: E0515 15:13:38.684173 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.684417 kubelet[2694]: W0515 15:13:38.684371 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.684417 kubelet[2694]: E0515 15:13:38.684396 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.687257 kubelet[2694]: E0515 15:13:38.687221 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:38.692186 kubelet[2694]: E0515 15:13:38.692145 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:38.693437 containerd[1549]: time="2025-05-15T15:13:38.693382695Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-qjmm6,Uid:06424021-71ea-4cef-bbec-94f6e5d79ae3,Namespace:calico-system,Attempt:0,}" May 15 15:13:38.704702 kubelet[2694]: E0515 15:13:38.704658 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.704702 kubelet[2694]: W0515 15:13:38.704690 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.705058 kubelet[2694]: E0515 15:13:38.704720 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.751980 kubelet[2694]: E0515 15:13:38.749146 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.751980 kubelet[2694]: W0515 15:13:38.749172 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.751980 kubelet[2694]: E0515 15:13:38.749201 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.751980 kubelet[2694]: E0515 15:13:38.750098 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.751980 kubelet[2694]: W0515 15:13:38.750116 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.751980 kubelet[2694]: E0515 15:13:38.751158 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.753006 kubelet[2694]: E0515 15:13:38.752974 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.753006 kubelet[2694]: W0515 15:13:38.752996 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.753947 kubelet[2694]: E0515 15:13:38.753018 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.755152 kubelet[2694]: E0515 15:13:38.755127 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.755152 kubelet[2694]: W0515 15:13:38.755147 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.755296 kubelet[2694]: E0515 15:13:38.755168 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.756401 kubelet[2694]: E0515 15:13:38.756381 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:38.756401 kubelet[2694]: W0515 15:13:38.756402 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:38.756639 kubelet[2694]: E0515 15:13:38.756421 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:38.761220 containerd[1549]: time="2025-05-15T15:13:38.760910800Z" level=info msg="connecting to shim 887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82" address="unix:///run/containerd/s/bc46e7a4428a679462b18c2d1cb7901b190dc09c18f931c75b7f5148f6288437" namespace=k8s.io protocol=ttrpc version=3 May 15 15:13:38.772658 containerd[1549]: time="2025-05-15T15:13:38.772112069Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-596f4965b7-vjc2l,Uid:12abe38c-3744-42fb-bd9c-1466a933b3e0,Namespace:calico-system,Attempt:0,} returns sandbox id \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\"" May 15 15:13:38.773253 kubelet[2694]: E0515 15:13:38.773222 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:38.775326 containerd[1549]: time="2025-05-15T15:13:38.775294567Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" May 15 15:13:38.813203 systemd[1]: Started cri-containerd-887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82.scope - libcontainer container 887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82. May 15 15:13:38.875208 containerd[1549]: time="2025-05-15T15:13:38.875153927Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-qjmm6,Uid:06424021-71ea-4cef-bbec-94f6e5d79ae3,Namespace:calico-system,Attempt:0,} returns sandbox id \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\"" May 15 15:13:38.877744 kubelet[2694]: E0515 15:13:38.877690 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:39.595794 kubelet[2694]: E0515 15:13:39.595360 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:39.662445 kubelet[2694]: E0515 15:13:39.662212 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:39.662445 kubelet[2694]: W0515 15:13:39.662248 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:39.662445 kubelet[2694]: E0515 15:13:39.662283 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:39.662916 kubelet[2694]: E0515 15:13:39.662893 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:39.663043 kubelet[2694]: W0515 15:13:39.663020 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:39.663359 kubelet[2694]: E0515 15:13:39.663201 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:39.663524 kubelet[2694]: E0515 15:13:39.663508 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:39.663612 kubelet[2694]: W0515 15:13:39.663593 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:39.663832 kubelet[2694]: E0515 15:13:39.663678 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:39.664046 kubelet[2694]: E0515 15:13:39.664029 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:39.664305 kubelet[2694]: W0515 15:13:39.664122 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:39.664305 kubelet[2694]: E0515 15:13:39.664145 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:39.664528 kubelet[2694]: E0515 15:13:39.664509 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:39.664630 kubelet[2694]: W0515 15:13:39.664612 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:39.664714 kubelet[2694]: E0515 15:13:39.664698 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:40.482434 kubelet[2694]: E0515 15:13:40.482252 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5mw49" podUID="ea7342ed-f869-4f54-b67c-23b40f5f8961" May 15 15:13:41.357534 containerd[1549]: time="2025-05-15T15:13:41.357461579Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:41.358957 containerd[1549]: time="2025-05-15T15:13:41.358770534Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.3: active requests=0, bytes read=30426870" May 15 15:13:41.359407 containerd[1549]: time="2025-05-15T15:13:41.359370534Z" level=info msg="ImageCreate event name:\"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:41.362457 containerd[1549]: time="2025-05-15T15:13:41.362395898Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:41.362773 containerd[1549]: time="2025-05-15T15:13:41.362745159Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.3\" with image id \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\", size \"31919484\" in 2.587199455s" May 15 15:13:41.362845 containerd[1549]: time="2025-05-15T15:13:41.362779243Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\"" May 15 15:13:41.364368 containerd[1549]: time="2025-05-15T15:13:41.364332241Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" May 15 15:13:41.382692 containerd[1549]: time="2025-05-15T15:13:41.382647809Z" level=info msg="CreateContainer within sandbox \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" May 15 15:13:41.389670 containerd[1549]: time="2025-05-15T15:13:41.388857511Z" level=info msg="Container c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:41.400057 containerd[1549]: time="2025-05-15T15:13:41.399957722Z" level=info msg="CreateContainer within sandbox \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\"" May 15 15:13:41.402699 containerd[1549]: time="2025-05-15T15:13:41.402474227Z" level=info msg="StartContainer for \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\"" May 15 15:13:41.406082 containerd[1549]: time="2025-05-15T15:13:41.406050695Z" level=info msg="connecting to shim c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5" address="unix:///run/containerd/s/757fbf1bcad6294ce35343044f403e61e29c683483b74643c9d8783cc603dd74" protocol=ttrpc version=3 May 15 15:13:41.434163 systemd[1]: Started cri-containerd-c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5.scope - libcontainer container c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5. May 15 15:13:41.503306 containerd[1549]: time="2025-05-15T15:13:41.503227372Z" level=info msg="StartContainer for \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" returns successfully" May 15 15:13:41.576356 update_engine[1528]: I20250515 15:13:41.575027 1528 update_attempter.cc:509] Updating boot flags... May 15 15:13:41.620409 kubelet[2694]: E0515 15:13:41.619423 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:41.683180 kubelet[2694]: E0515 15:13:41.683082 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.683180 kubelet[2694]: W0515 15:13:41.683113 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.683552 kubelet[2694]: E0515 15:13:41.683389 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.684157 kubelet[2694]: E0515 15:13:41.684144 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.684351 kubelet[2694]: W0515 15:13:41.684235 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.684351 kubelet[2694]: E0515 15:13:41.684255 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.688818 kubelet[2694]: E0515 15:13:41.688491 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.688818 kubelet[2694]: W0515 15:13:41.688516 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.688818 kubelet[2694]: E0515 15:13:41.688538 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.690356 kubelet[2694]: E0515 15:13:41.690324 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.690356 kubelet[2694]: W0515 15:13:41.690340 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.690778 kubelet[2694]: E0515 15:13:41.690477 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.691524 kubelet[2694]: E0515 15:13:41.691432 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.691524 kubelet[2694]: W0515 15:13:41.691453 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.692017 kubelet[2694]: E0515 15:13:41.691734 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.692623 kubelet[2694]: E0515 15:13:41.692490 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.692623 kubelet[2694]: W0515 15:13:41.692505 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.692623 kubelet[2694]: E0515 15:13:41.692517 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.693840 kubelet[2694]: E0515 15:13:41.693774 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.693840 kubelet[2694]: W0515 15:13:41.693788 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.694533 kubelet[2694]: E0515 15:13:41.693802 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.695803 kubelet[2694]: E0515 15:13:41.695684 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.695803 kubelet[2694]: W0515 15:13:41.695698 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.695803 kubelet[2694]: E0515 15:13:41.695710 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.697075 kubelet[2694]: E0515 15:13:41.696730 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.697075 kubelet[2694]: W0515 15:13:41.696747 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.697075 kubelet[2694]: E0515 15:13:41.696762 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.697710 kubelet[2694]: E0515 15:13:41.697678 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.697865 kubelet[2694]: W0515 15:13:41.697695 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.697865 kubelet[2694]: E0515 15:13:41.697810 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.699671 kubelet[2694]: E0515 15:13:41.699329 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.700302 kubelet[2694]: W0515 15:13:41.699620 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.700978 kubelet[2694]: E0515 15:13:41.700961 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.701517 kubelet[2694]: E0515 15:13:41.701464 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.701954 kubelet[2694]: W0515 15:13:41.701781 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.701954 kubelet[2694]: E0515 15:13:41.701804 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.702421 kubelet[2694]: E0515 15:13:41.702396 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.702534 kubelet[2694]: W0515 15:13:41.702409 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.702534 kubelet[2694]: E0515 15:13:41.702495 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.702813 kubelet[2694]: E0515 15:13:41.702782 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.702813 kubelet[2694]: W0515 15:13:41.702795 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.702954 kubelet[2694]: E0515 15:13:41.702893 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.703273 kubelet[2694]: E0515 15:13:41.703240 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.703402 kubelet[2694]: W0515 15:13:41.703336 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.703402 kubelet[2694]: E0515 15:13:41.703356 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.703988 kubelet[2694]: E0515 15:13:41.703965 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.704118 kubelet[2694]: W0515 15:13:41.704104 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.704316 kubelet[2694]: E0515 15:13:41.704253 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.704859 kubelet[2694]: E0515 15:13:41.704844 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.705087 kubelet[2694]: W0515 15:13:41.704970 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.705087 kubelet[2694]: E0515 15:13:41.704990 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.706170 kubelet[2694]: E0515 15:13:41.706151 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.706398 kubelet[2694]: W0515 15:13:41.706223 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.706398 kubelet[2694]: E0515 15:13:41.706239 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.706702 kubelet[2694]: E0515 15:13:41.706667 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.706702 kubelet[2694]: W0515 15:13:41.706687 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.706960 kubelet[2694]: E0515 15:13:41.706801 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.707378 kubelet[2694]: E0515 15:13:41.707365 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.707488 kubelet[2694]: W0515 15:13:41.707407 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.707674 kubelet[2694]: E0515 15:13:41.707560 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.708385 kubelet[2694]: E0515 15:13:41.708345 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.708385 kubelet[2694]: W0515 15:13:41.708358 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.708594 kubelet[2694]: E0515 15:13:41.708371 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.710049 kubelet[2694]: E0515 15:13:41.710006 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.710049 kubelet[2694]: W0515 15:13:41.710020 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.710346 kubelet[2694]: E0515 15:13:41.710035 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.711046 kubelet[2694]: E0515 15:13:41.710997 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.711046 kubelet[2694]: W0515 15:13:41.711012 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.711046 kubelet[2694]: E0515 15:13:41.711028 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.712167 kubelet[2694]: E0515 15:13:41.711979 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.712167 kubelet[2694]: W0515 15:13:41.712054 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.712167 kubelet[2694]: E0515 15:13:41.712068 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.712519 kubelet[2694]: E0515 15:13:41.712478 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.712666 kubelet[2694]: W0515 15:13:41.712578 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.712666 kubelet[2694]: E0515 15:13:41.712598 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.713217 kubelet[2694]: E0515 15:13:41.713119 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.713217 kubelet[2694]: W0515 15:13:41.713133 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.713217 kubelet[2694]: E0515 15:13:41.713144 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.713472 kubelet[2694]: E0515 15:13:41.713428 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.713472 kubelet[2694]: W0515 15:13:41.713439 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.713472 kubelet[2694]: E0515 15:13:41.713451 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.713962 kubelet[2694]: E0515 15:13:41.713910 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.714253 kubelet[2694]: W0515 15:13:41.713924 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.714253 kubelet[2694]: E0515 15:13:41.714204 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.716270 kubelet[2694]: E0515 15:13:41.716239 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.716504 kubelet[2694]: W0515 15:13:41.716378 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.716504 kubelet[2694]: E0515 15:13:41.716404 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.716749 kubelet[2694]: E0515 15:13:41.716711 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.716749 kubelet[2694]: W0515 15:13:41.716725 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.716916 kubelet[2694]: E0515 15:13:41.716738 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.717216 kubelet[2694]: E0515 15:13:41.717170 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.717216 kubelet[2694]: W0515 15:13:41.717183 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.717216 kubelet[2694]: E0515 15:13:41.717196 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.718853 kubelet[2694]: E0515 15:13:41.718782 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.718853 kubelet[2694]: W0515 15:13:41.718799 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.718853 kubelet[2694]: E0515 15:13:41.718814 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:41.719727 kubelet[2694]: E0515 15:13:41.719687 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:41.719727 kubelet[2694]: W0515 15:13:41.719701 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:41.720026 kubelet[2694]: E0515 15:13:41.719714 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.482340 kubelet[2694]: E0515 15:13:42.482217 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5mw49" podUID="ea7342ed-f869-4f54-b67c-23b40f5f8961" May 15 15:13:42.621821 kubelet[2694]: I0515 15:13:42.621793 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:13:42.622683 kubelet[2694]: E0515 15:13:42.622663 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:42.709180 kubelet[2694]: E0515 15:13:42.709145 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.709180 kubelet[2694]: W0515 15:13:42.709178 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.709395 kubelet[2694]: E0515 15:13:42.709201 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.709428 kubelet[2694]: E0515 15:13:42.709410 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.709428 kubelet[2694]: W0515 15:13:42.709420 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.709480 kubelet[2694]: E0515 15:13:42.709432 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.710045 kubelet[2694]: E0515 15:13:42.709594 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.710045 kubelet[2694]: W0515 15:13:42.709606 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.710045 kubelet[2694]: E0515 15:13:42.709615 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.710045 kubelet[2694]: E0515 15:13:42.709780 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.710045 kubelet[2694]: W0515 15:13:42.709786 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.710045 kubelet[2694]: E0515 15:13:42.709793 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.710045 kubelet[2694]: E0515 15:13:42.709963 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.710045 kubelet[2694]: W0515 15:13:42.709975 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.710045 kubelet[2694]: E0515 15:13:42.709986 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711257 kubelet[2694]: E0515 15:13:42.710193 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711257 kubelet[2694]: W0515 15:13:42.710202 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711257 kubelet[2694]: E0515 15:13:42.710212 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711257 kubelet[2694]: E0515 15:13:42.710375 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711257 kubelet[2694]: W0515 15:13:42.710382 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711257 kubelet[2694]: E0515 15:13:42.710403 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711257 kubelet[2694]: E0515 15:13:42.710579 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711257 kubelet[2694]: W0515 15:13:42.710589 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711257 kubelet[2694]: E0515 15:13:42.710601 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711257 kubelet[2694]: E0515 15:13:42.710834 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711558 kubelet[2694]: W0515 15:13:42.710843 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711558 kubelet[2694]: E0515 15:13:42.710856 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711558 kubelet[2694]: E0515 15:13:42.711147 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711558 kubelet[2694]: W0515 15:13:42.711157 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711558 kubelet[2694]: E0515 15:13:42.711167 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711558 kubelet[2694]: E0515 15:13:42.711333 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711558 kubelet[2694]: W0515 15:13:42.711341 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711558 kubelet[2694]: E0515 15:13:42.711350 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711558 kubelet[2694]: E0515 15:13:42.711471 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711558 kubelet[2694]: W0515 15:13:42.711477 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711802 kubelet[2694]: E0515 15:13:42.711483 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711802 kubelet[2694]: E0515 15:13:42.711608 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711802 kubelet[2694]: W0515 15:13:42.711613 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711802 kubelet[2694]: E0515 15:13:42.711639 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.711802 kubelet[2694]: E0515 15:13:42.711760 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.711802 kubelet[2694]: W0515 15:13:42.711766 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.711802 kubelet[2694]: E0515 15:13:42.711774 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.712817 kubelet[2694]: E0515 15:13:42.711890 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.712817 kubelet[2694]: W0515 15:13:42.711896 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.712817 kubelet[2694]: E0515 15:13:42.711902 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.718040 kubelet[2694]: E0515 15:13:42.717965 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.718040 kubelet[2694]: W0515 15:13:42.717990 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.718040 kubelet[2694]: E0515 15:13:42.718013 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.718560 kubelet[2694]: E0515 15:13:42.718521 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.718560 kubelet[2694]: W0515 15:13:42.718534 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.718886 kubelet[2694]: E0515 15:13:42.718807 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.719028 kubelet[2694]: E0515 15:13:42.719018 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.719121 kubelet[2694]: W0515 15:13:42.719098 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.719179 kubelet[2694]: E0515 15:13:42.719170 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.719372 kubelet[2694]: E0515 15:13:42.719353 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.719372 kubelet[2694]: W0515 15:13:42.719370 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.719561 kubelet[2694]: E0515 15:13:42.719389 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.719664 kubelet[2694]: E0515 15:13:42.719653 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.719732 kubelet[2694]: W0515 15:13:42.719722 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.719844 kubelet[2694]: E0515 15:13:42.719781 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.720092 kubelet[2694]: E0515 15:13:42.720079 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.720290 kubelet[2694]: W0515 15:13:42.720180 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.720290 kubelet[2694]: E0515 15:13:42.720204 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.720429 kubelet[2694]: E0515 15:13:42.720420 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.720533 kubelet[2694]: W0515 15:13:42.720462 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.720533 kubelet[2694]: E0515 15:13:42.720488 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.720770 kubelet[2694]: E0515 15:13:42.720750 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.720881 kubelet[2694]: W0515 15:13:42.720761 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.720881 kubelet[2694]: E0515 15:13:42.720852 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.721140 kubelet[2694]: E0515 15:13:42.721116 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.721140 kubelet[2694]: W0515 15:13:42.721128 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.721375 kubelet[2694]: E0515 15:13:42.721247 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.721492 kubelet[2694]: E0515 15:13:42.721482 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.721567 kubelet[2694]: W0515 15:13:42.721557 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.721702 kubelet[2694]: E0515 15:13:42.721615 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.721841 kubelet[2694]: E0515 15:13:42.721828 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.721877 kubelet[2694]: W0515 15:13:42.721841 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.721877 kubelet[2694]: E0515 15:13:42.721852 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.722051 kubelet[2694]: E0515 15:13:42.722032 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.722051 kubelet[2694]: W0515 15:13:42.722042 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.722051 kubelet[2694]: E0515 15:13:42.722050 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.722340 kubelet[2694]: E0515 15:13:42.722326 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.722340 kubelet[2694]: W0515 15:13:42.722340 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.722409 kubelet[2694]: E0515 15:13:42.722350 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.722502 kubelet[2694]: E0515 15:13:42.722490 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.722502 kubelet[2694]: W0515 15:13:42.722501 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.722576 kubelet[2694]: E0515 15:13:42.722508 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.722698 kubelet[2694]: E0515 15:13:42.722627 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.722698 kubelet[2694]: W0515 15:13:42.722642 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.722698 kubelet[2694]: E0515 15:13:42.722653 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.722975 kubelet[2694]: E0515 15:13:42.722962 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.723117 kubelet[2694]: W0515 15:13:42.723036 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.723117 kubelet[2694]: E0515 15:13:42.723065 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.723601 kubelet[2694]: E0515 15:13:42.723339 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.723601 kubelet[2694]: W0515 15:13:42.723350 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.723601 kubelet[2694]: E0515 15:13:42.723360 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:42.723757 kubelet[2694]: E0515 15:13:42.723747 2694 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 15:13:42.723815 kubelet[2694]: W0515 15:13:42.723806 2694 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 15:13:42.723868 kubelet[2694]: E0515 15:13:42.723857 2694 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 15:13:44.029173 containerd[1549]: time="2025-05-15T15:13:44.029119995Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:44.030276 containerd[1549]: time="2025-05-15T15:13:44.030063953Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3: active requests=0, bytes read=5366937" May 15 15:13:44.030955 containerd[1549]: time="2025-05-15T15:13:44.030842363Z" level=info msg="ImageCreate event name:\"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:44.032622 containerd[1549]: time="2025-05-15T15:13:44.032585462Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:44.033653 containerd[1549]: time="2025-05-15T15:13:44.033517269Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" with image id \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\", size \"6859519\" in 2.669149966s" May 15 15:13:44.033653 containerd[1549]: time="2025-05-15T15:13:44.033556973Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\"" May 15 15:13:44.036397 containerd[1549]: time="2025-05-15T15:13:44.036357645Z" level=info msg="CreateContainer within sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 15 15:13:44.047958 containerd[1549]: time="2025-05-15T15:13:44.046152311Z" level=info msg="Container 5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:44.051746 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2006779642.mount: Deactivated successfully. May 15 15:13:44.061327 containerd[1549]: time="2025-05-15T15:13:44.061135425Z" level=info msg="CreateContainer within sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\"" May 15 15:13:44.064024 containerd[1549]: time="2025-05-15T15:13:44.063115981Z" level=info msg="StartContainer for \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\"" May 15 15:13:44.065408 containerd[1549]: time="2025-05-15T15:13:44.065368221Z" level=info msg="connecting to shim 5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a" address="unix:///run/containerd/s/bc46e7a4428a679462b18c2d1cb7901b190dc09c18f931c75b7f5148f6288437" protocol=ttrpc version=3 May 15 15:13:44.094206 systemd[1]: Started cri-containerd-5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a.scope - libcontainer container 5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a. May 15 15:13:44.160562 containerd[1549]: time="2025-05-15T15:13:44.160493672Z" level=info msg="StartContainer for \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\" returns successfully" May 15 15:13:44.178942 systemd[1]: cri-containerd-5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a.scope: Deactivated successfully. May 15 15:13:44.203468 containerd[1549]: time="2025-05-15T15:13:44.203401551Z" level=info msg="received exit event container_id:\"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\" id:\"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\" pid:3372 exited_at:{seconds:1747322024 nanos:181798148}" May 15 15:13:44.205480 containerd[1549]: time="2025-05-15T15:13:44.205420961Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\" id:\"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\" pid:3372 exited_at:{seconds:1747322024 nanos:181798148}" May 15 15:13:44.238668 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a-rootfs.mount: Deactivated successfully. May 15 15:13:44.483154 kubelet[2694]: E0515 15:13:44.482777 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5mw49" podUID="ea7342ed-f869-4f54-b67c-23b40f5f8961" May 15 15:13:44.630220 kubelet[2694]: E0515 15:13:44.630184 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:44.632213 containerd[1549]: time="2025-05-15T15:13:44.632109760Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" May 15 15:13:44.655711 kubelet[2694]: I0515 15:13:44.655651 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-596f4965b7-vjc2l" podStartSLOduration=4.066452364 podStartE2EDuration="6.655631902s" podCreationTimestamp="2025-05-15 15:13:38 +0000 UTC" firstStartedPulling="2025-05-15 15:13:38.774714402 +0000 UTC m=+12.455587409" lastFinishedPulling="2025-05-15 15:13:41.363893929 +0000 UTC m=+15.044766947" observedRunningTime="2025-05-15 15:13:41.645821195 +0000 UTC m=+15.326694241" watchObservedRunningTime="2025-05-15 15:13:44.655631902 +0000 UTC m=+18.336504934" May 15 15:13:46.483661 kubelet[2694]: E0515 15:13:46.483238 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5mw49" podUID="ea7342ed-f869-4f54-b67c-23b40f5f8961" May 15 15:13:48.482902 kubelet[2694]: E0515 15:13:48.482850 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5mw49" podUID="ea7342ed-f869-4f54-b67c-23b40f5f8961" May 15 15:13:49.212510 systemd[1]: Started sshd@8-24.144.90.91:22-60.167.166.161:56126.service - OpenSSH per-connection server daemon (60.167.166.161:56126). May 15 15:13:50.038311 containerd[1549]: time="2025-05-15T15:13:50.036974159Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:50.038311 containerd[1549]: time="2025-05-15T15:13:50.037693223Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.3: active requests=0, bytes read=97793683" May 15 15:13:50.038311 containerd[1549]: time="2025-05-15T15:13:50.037766532Z" level=info msg="ImageCreate event name:\"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:50.039599 containerd[1549]: time="2025-05-15T15:13:50.039566782Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:50.040622 containerd[1549]: time="2025-05-15T15:13:50.040562654Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.3\" with image id \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\", size \"99286305\" in 5.408354221s" May 15 15:13:50.040622 containerd[1549]: time="2025-05-15T15:13:50.040610489Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\"" May 15 15:13:50.047080 containerd[1549]: time="2025-05-15T15:13:50.045498351Z" level=info msg="CreateContainer within sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 15 15:13:50.057037 containerd[1549]: time="2025-05-15T15:13:50.056988028Z" level=info msg="Container 642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:50.063176 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount334360133.mount: Deactivated successfully. May 15 15:13:50.072987 containerd[1549]: time="2025-05-15T15:13:50.072901286Z" level=info msg="CreateContainer within sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\"" May 15 15:13:50.073814 containerd[1549]: time="2025-05-15T15:13:50.073749256Z" level=info msg="StartContainer for \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\"" May 15 15:13:50.076340 containerd[1549]: time="2025-05-15T15:13:50.076260775Z" level=info msg="connecting to shim 642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c" address="unix:///run/containerd/s/bc46e7a4428a679462b18c2d1cb7901b190dc09c18f931c75b7f5148f6288437" protocol=ttrpc version=3 May 15 15:13:50.107320 systemd[1]: Started cri-containerd-642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c.scope - libcontainer container 642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c. May 15 15:13:50.155207 containerd[1549]: time="2025-05-15T15:13:50.155168089Z" level=info msg="StartContainer for \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\" returns successfully" May 15 15:13:50.482723 kubelet[2694]: E0515 15:13:50.482656 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5mw49" podUID="ea7342ed-f869-4f54-b67c-23b40f5f8961" May 15 15:13:50.673127 systemd[1]: cri-containerd-642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c.scope: Deactivated successfully. May 15 15:13:50.673510 systemd[1]: cri-containerd-642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c.scope: Consumed 530ms CPU time, 148.6M memory peak, 4.3M read from disk, 154M written to disk. May 15 15:13:50.699111 containerd[1549]: time="2025-05-15T15:13:50.699061026Z" level=info msg="received exit event container_id:\"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\" id:\"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\" pid:3433 exited_at:{seconds:1747322030 nanos:675907052}" May 15 15:13:50.701198 containerd[1549]: time="2025-05-15T15:13:50.701063015Z" level=info msg="TaskExit event in podsandbox handler container_id:\"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\" id:\"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\" pid:3433 exited_at:{seconds:1747322030 nanos:675907052}" May 15 15:13:50.716441 kubelet[2694]: E0515 15:13:50.716361 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:50.756503 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c-rootfs.mount: Deactivated successfully. May 15 15:13:50.815635 kubelet[2694]: I0515 15:13:50.815590 2694 kubelet_node_status.go:502] "Fast updating node status as it just became ready" May 15 15:13:50.868701 systemd[1]: Created slice kubepods-burstable-pod64abed57_9eb1_4790_953d_d2c240a13afe.slice - libcontainer container kubepods-burstable-pod64abed57_9eb1_4790_953d_d2c240a13afe.slice. May 15 15:13:50.886397 kubelet[2694]: I0515 15:13:50.886347 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7ffg\" (UniqueName: \"kubernetes.io/projected/64abed57-9eb1-4790-953d-d2c240a13afe-kube-api-access-c7ffg\") pod \"coredns-668d6bf9bc-vrxcl\" (UID: \"64abed57-9eb1-4790-953d-d2c240a13afe\") " pod="kube-system/coredns-668d6bf9bc-vrxcl" May 15 15:13:50.886397 kubelet[2694]: I0515 15:13:50.886406 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64abed57-9eb1-4790-953d-d2c240a13afe-config-volume\") pod \"coredns-668d6bf9bc-vrxcl\" (UID: \"64abed57-9eb1-4790-953d-d2c240a13afe\") " pod="kube-system/coredns-668d6bf9bc-vrxcl" May 15 15:13:50.890164 systemd[1]: Created slice kubepods-besteffort-pod826ead8b_3406_4df0_8588_b2656fcca004.slice - libcontainer container kubepods-besteffort-pod826ead8b_3406_4df0_8588_b2656fcca004.slice. May 15 15:13:50.917066 systemd[1]: Created slice kubepods-burstable-podd13c79a0_d35b_4259_aebf_ab7b67e15ff2.slice - libcontainer container kubepods-burstable-podd13c79a0_d35b_4259_aebf_ab7b67e15ff2.slice. May 15 15:13:50.924010 systemd[1]: Created slice kubepods-besteffort-pod52c8dc92_1a03_4e1d_bc5b_3844d9a477ad.slice - libcontainer container kubepods-besteffort-pod52c8dc92_1a03_4e1d_bc5b_3844d9a477ad.slice. May 15 15:13:50.933407 systemd[1]: Created slice kubepods-besteffort-pod1c80ddf0_e33c_4576_8bd9_03a8cc5f33e7.slice - libcontainer container kubepods-besteffort-pod1c80ddf0_e33c_4576_8bd9_03a8cc5f33e7.slice. May 15 15:13:50.950419 systemd[1]: Created slice kubepods-besteffort-pod869a7900_ca01_45de_9a8f_19a49b2d754b.slice - libcontainer container kubepods-besteffort-pod869a7900_ca01_45de_9a8f_19a49b2d754b.slice. May 15 15:13:50.987167 kubelet[2694]: I0515 15:13:50.987097 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d13c79a0-d35b-4259-aebf-ab7b67e15ff2-config-volume\") pod \"coredns-668d6bf9bc-5gp9s\" (UID: \"d13c79a0-d35b-4259-aebf-ab7b67e15ff2\") " pod="kube-system/coredns-668d6bf9bc-5gp9s" May 15 15:13:50.987339 kubelet[2694]: I0515 15:13:50.987185 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrs64\" (UniqueName: \"kubernetes.io/projected/826ead8b-3406-4df0-8588-b2656fcca004-kube-api-access-xrs64\") pod \"calico-kube-controllers-c8cd4b68c-85hmc\" (UID: \"826ead8b-3406-4df0-8588-b2656fcca004\") " pod="calico-system/calico-kube-controllers-c8cd4b68c-85hmc" May 15 15:13:50.987339 kubelet[2694]: I0515 15:13:50.987238 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7-calico-apiserver-certs\") pod \"calico-apiserver-585b9b4df4-d2bqf\" (UID: \"1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7\") " pod="calico-apiserver/calico-apiserver-585b9b4df4-d2bqf" May 15 15:13:50.987339 kubelet[2694]: I0515 15:13:50.987307 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktx9g\" (UniqueName: \"kubernetes.io/projected/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad-kube-api-access-ktx9g\") pod \"calico-apiserver-585b9b4df4-cr2q4\" (UID: \"52c8dc92-1a03-4e1d-bc5b-3844d9a477ad\") " pod="calico-apiserver/calico-apiserver-585b9b4df4-cr2q4" May 15 15:13:50.987445 kubelet[2694]: I0515 15:13:50.987352 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6w9f\" (UniqueName: \"kubernetes.io/projected/d13c79a0-d35b-4259-aebf-ab7b67e15ff2-kube-api-access-b6w9f\") pod \"coredns-668d6bf9bc-5gp9s\" (UID: \"d13c79a0-d35b-4259-aebf-ab7b67e15ff2\") " pod="kube-system/coredns-668d6bf9bc-5gp9s" May 15 15:13:50.987445 kubelet[2694]: I0515 15:13:50.987383 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw7r8\" (UniqueName: \"kubernetes.io/projected/869a7900-ca01-45de-9a8f-19a49b2d754b-kube-api-access-mw7r8\") pod \"calico-apiserver-6999544d4f-js9sf\" (UID: \"869a7900-ca01-45de-9a8f-19a49b2d754b\") " pod="calico-apiserver/calico-apiserver-6999544d4f-js9sf" May 15 15:13:50.987445 kubelet[2694]: I0515 15:13:50.987400 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2brf\" (UniqueName: \"kubernetes.io/projected/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7-kube-api-access-q2brf\") pod \"calico-apiserver-585b9b4df4-d2bqf\" (UID: \"1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7\") " pod="calico-apiserver/calico-apiserver-585b9b4df4-d2bqf" May 15 15:13:50.987445 kubelet[2694]: I0515 15:13:50.987428 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad-calico-apiserver-certs\") pod \"calico-apiserver-585b9b4df4-cr2q4\" (UID: \"52c8dc92-1a03-4e1d-bc5b-3844d9a477ad\") " pod="calico-apiserver/calico-apiserver-585b9b4df4-cr2q4" May 15 15:13:50.987550 kubelet[2694]: I0515 15:13:50.987456 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/869a7900-ca01-45de-9a8f-19a49b2d754b-calico-apiserver-certs\") pod \"calico-apiserver-6999544d4f-js9sf\" (UID: \"869a7900-ca01-45de-9a8f-19a49b2d754b\") " pod="calico-apiserver/calico-apiserver-6999544d4f-js9sf" May 15 15:13:50.987550 kubelet[2694]: I0515 15:13:50.987476 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/826ead8b-3406-4df0-8588-b2656fcca004-tigera-ca-bundle\") pod \"calico-kube-controllers-c8cd4b68c-85hmc\" (UID: \"826ead8b-3406-4df0-8588-b2656fcca004\") " pod="calico-system/calico-kube-controllers-c8cd4b68c-85hmc" May 15 15:13:51.178052 kubelet[2694]: E0515 15:13:51.177811 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:51.179669 containerd[1549]: time="2025-05-15T15:13:51.179627899Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vrxcl,Uid:64abed57-9eb1-4790-953d-d2c240a13afe,Namespace:kube-system,Attempt:0,}" May 15 15:13:51.215026 containerd[1549]: time="2025-05-15T15:13:51.214913957Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-c8cd4b68c-85hmc,Uid:826ead8b-3406-4df0-8588-b2656fcca004,Namespace:calico-system,Attempt:0,}" May 15 15:13:51.221171 kubelet[2694]: E0515 15:13:51.221127 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:51.226793 containerd[1549]: time="2025-05-15T15:13:51.226731146Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5gp9s,Uid:d13c79a0-d35b-4259-aebf-ab7b67e15ff2,Namespace:kube-system,Attempt:0,}" May 15 15:13:51.230476 containerd[1549]: time="2025-05-15T15:13:51.230410775Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-585b9b4df4-cr2q4,Uid:52c8dc92-1a03-4e1d-bc5b-3844d9a477ad,Namespace:calico-apiserver,Attempt:0,}" May 15 15:13:51.255026 containerd[1549]: time="2025-05-15T15:13:51.254912183Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-585b9b4df4-d2bqf,Uid:1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7,Namespace:calico-apiserver,Attempt:0,}" May 15 15:13:51.257781 containerd[1549]: time="2025-05-15T15:13:51.257560301Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6999544d4f-js9sf,Uid:869a7900-ca01-45de-9a8f-19a49b2d754b,Namespace:calico-apiserver,Attempt:0,}" May 15 15:13:51.465876 containerd[1549]: time="2025-05-15T15:13:51.465801158Z" level=error msg="Failed to destroy network for sandbox \"1107c4c6190f8a8a283dc86f88edbef23c49f5e8d528479e6ba7591f8e77b006\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.495099 containerd[1549]: time="2025-05-15T15:13:51.495030777Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-585b9b4df4-d2bqf,Uid:1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1107c4c6190f8a8a283dc86f88edbef23c49f5e8d528479e6ba7591f8e77b006\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.495793 kubelet[2694]: E0515 15:13:51.495590 2694 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1107c4c6190f8a8a283dc86f88edbef23c49f5e8d528479e6ba7591f8e77b006\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.497095 kubelet[2694]: E0515 15:13:51.496401 2694 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1107c4c6190f8a8a283dc86f88edbef23c49f5e8d528479e6ba7591f8e77b006\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-585b9b4df4-d2bqf" May 15 15:13:51.497095 kubelet[2694]: E0515 15:13:51.496718 2694 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1107c4c6190f8a8a283dc86f88edbef23c49f5e8d528479e6ba7591f8e77b006\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-585b9b4df4-d2bqf" May 15 15:13:51.498029 kubelet[2694]: E0515 15:13:51.497907 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-585b9b4df4-d2bqf_calico-apiserver(1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-585b9b4df4-d2bqf_calico-apiserver(1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1107c4c6190f8a8a283dc86f88edbef23c49f5e8d528479e6ba7591f8e77b006\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-585b9b4df4-d2bqf" podUID="1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7" May 15 15:13:51.501306 containerd[1549]: time="2025-05-15T15:13:51.501146640Z" level=error msg="Failed to destroy network for sandbox \"41c0c0859c8e917038957d2df90e023103066046406870e4942e89a9f9bffdb4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.502957 containerd[1549]: time="2025-05-15T15:13:51.502804801Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vrxcl,Uid:64abed57-9eb1-4790-953d-d2c240a13afe,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"41c0c0859c8e917038957d2df90e023103066046406870e4942e89a9f9bffdb4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.503340 containerd[1549]: time="2025-05-15T15:13:51.502999635Z" level=error msg="Failed to destroy network for sandbox \"9f752a70a0339da6910b25392c596e1ef9cc6e6eb0502840674b4e36e614cee8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.503572 kubelet[2694]: E0515 15:13:51.503360 2694 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41c0c0859c8e917038957d2df90e023103066046406870e4942e89a9f9bffdb4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.503572 kubelet[2694]: E0515 15:13:51.503420 2694 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41c0c0859c8e917038957d2df90e023103066046406870e4942e89a9f9bffdb4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vrxcl" May 15 15:13:51.503572 kubelet[2694]: E0515 15:13:51.503471 2694 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41c0c0859c8e917038957d2df90e023103066046406870e4942e89a9f9bffdb4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vrxcl" May 15 15:13:51.505295 kubelet[2694]: E0515 15:13:51.503563 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-vrxcl_kube-system(64abed57-9eb1-4790-953d-d2c240a13afe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-vrxcl_kube-system(64abed57-9eb1-4790-953d-d2c240a13afe)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"41c0c0859c8e917038957d2df90e023103066046406870e4942e89a9f9bffdb4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-vrxcl" podUID="64abed57-9eb1-4790-953d-d2c240a13afe" May 15 15:13:51.507493 containerd[1549]: time="2025-05-15T15:13:51.507449027Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-585b9b4df4-cr2q4,Uid:52c8dc92-1a03-4e1d-bc5b-3844d9a477ad,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9f752a70a0339da6910b25392c596e1ef9cc6e6eb0502840674b4e36e614cee8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.508460 kubelet[2694]: E0515 15:13:51.507878 2694 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9f752a70a0339da6910b25392c596e1ef9cc6e6eb0502840674b4e36e614cee8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.508802 kubelet[2694]: E0515 15:13:51.508776 2694 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9f752a70a0339da6910b25392c596e1ef9cc6e6eb0502840674b4e36e614cee8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-585b9b4df4-cr2q4" May 15 15:13:51.508892 kubelet[2694]: E0515 15:13:51.508880 2694 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9f752a70a0339da6910b25392c596e1ef9cc6e6eb0502840674b4e36e614cee8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-585b9b4df4-cr2q4" May 15 15:13:51.509007 kubelet[2694]: E0515 15:13:51.508986 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-585b9b4df4-cr2q4_calico-apiserver(52c8dc92-1a03-4e1d-bc5b-3844d9a477ad)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-585b9b4df4-cr2q4_calico-apiserver(52c8dc92-1a03-4e1d-bc5b-3844d9a477ad)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9f752a70a0339da6910b25392c596e1ef9cc6e6eb0502840674b4e36e614cee8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-585b9b4df4-cr2q4" podUID="52c8dc92-1a03-4e1d-bc5b-3844d9a477ad" May 15 15:13:51.517649 containerd[1549]: time="2025-05-15T15:13:51.517596337Z" level=error msg="Failed to destroy network for sandbox \"b55879edf77fb67fc35a4f9417aba531d5660104a4d922945efd96a00ca07424\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.518539 containerd[1549]: time="2025-05-15T15:13:51.518479386Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6999544d4f-js9sf,Uid:869a7900-ca01-45de-9a8f-19a49b2d754b,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b55879edf77fb67fc35a4f9417aba531d5660104a4d922945efd96a00ca07424\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.519058 kubelet[2694]: E0515 15:13:51.518731 2694 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b55879edf77fb67fc35a4f9417aba531d5660104a4d922945efd96a00ca07424\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.519058 kubelet[2694]: E0515 15:13:51.518790 2694 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b55879edf77fb67fc35a4f9417aba531d5660104a4d922945efd96a00ca07424\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6999544d4f-js9sf" May 15 15:13:51.519058 kubelet[2694]: E0515 15:13:51.518826 2694 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b55879edf77fb67fc35a4f9417aba531d5660104a4d922945efd96a00ca07424\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6999544d4f-js9sf" May 15 15:13:51.519773 kubelet[2694]: E0515 15:13:51.519520 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6999544d4f-js9sf_calico-apiserver(869a7900-ca01-45de-9a8f-19a49b2d754b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6999544d4f-js9sf_calico-apiserver(869a7900-ca01-45de-9a8f-19a49b2d754b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b55879edf77fb67fc35a4f9417aba531d5660104a4d922945efd96a00ca07424\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6999544d4f-js9sf" podUID="869a7900-ca01-45de-9a8f-19a49b2d754b" May 15 15:13:51.528206 containerd[1549]: time="2025-05-15T15:13:51.528096444Z" level=error msg="Failed to destroy network for sandbox \"1c88235c73141ef2fedab58bd024bfdd3fc8380496e273a8b1a65048e9fbfe86\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.529452 containerd[1549]: time="2025-05-15T15:13:51.529398419Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-c8cd4b68c-85hmc,Uid:826ead8b-3406-4df0-8588-b2656fcca004,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c88235c73141ef2fedab58bd024bfdd3fc8380496e273a8b1a65048e9fbfe86\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.530046 kubelet[2694]: E0515 15:13:51.529913 2694 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c88235c73141ef2fedab58bd024bfdd3fc8380496e273a8b1a65048e9fbfe86\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.530401 kubelet[2694]: E0515 15:13:51.530194 2694 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c88235c73141ef2fedab58bd024bfdd3fc8380496e273a8b1a65048e9fbfe86\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-c8cd4b68c-85hmc" May 15 15:13:51.530401 kubelet[2694]: E0515 15:13:51.530221 2694 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c88235c73141ef2fedab58bd024bfdd3fc8380496e273a8b1a65048e9fbfe86\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-c8cd4b68c-85hmc" May 15 15:13:51.530495 containerd[1549]: time="2025-05-15T15:13:51.529999969Z" level=error msg="Failed to destroy network for sandbox \"78be5076acc5fd27bf1c8ccfa9b3bf55b66b0d147ea7730764d8cdbee7b95868\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.530746 kubelet[2694]: E0515 15:13:51.530273 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-c8cd4b68c-85hmc_calico-system(826ead8b-3406-4df0-8588-b2656fcca004)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-c8cd4b68c-85hmc_calico-system(826ead8b-3406-4df0-8588-b2656fcca004)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1c88235c73141ef2fedab58bd024bfdd3fc8380496e273a8b1a65048e9fbfe86\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-c8cd4b68c-85hmc" podUID="826ead8b-3406-4df0-8588-b2656fcca004" May 15 15:13:51.531283 containerd[1549]: time="2025-05-15T15:13:51.531240436Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5gp9s,Uid:d13c79a0-d35b-4259-aebf-ab7b67e15ff2,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"78be5076acc5fd27bf1c8ccfa9b3bf55b66b0d147ea7730764d8cdbee7b95868\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.532124 kubelet[2694]: E0515 15:13:51.532085 2694 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"78be5076acc5fd27bf1c8ccfa9b3bf55b66b0d147ea7730764d8cdbee7b95868\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:51.532206 kubelet[2694]: E0515 15:13:51.532159 2694 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"78be5076acc5fd27bf1c8ccfa9b3bf55b66b0d147ea7730764d8cdbee7b95868\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-5gp9s" May 15 15:13:51.532206 kubelet[2694]: E0515 15:13:51.532182 2694 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"78be5076acc5fd27bf1c8ccfa9b3bf55b66b0d147ea7730764d8cdbee7b95868\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-5gp9s" May 15 15:13:51.532328 kubelet[2694]: E0515 15:13:51.532244 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-5gp9s_kube-system(d13c79a0-d35b-4259-aebf-ab7b67e15ff2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-5gp9s_kube-system(d13c79a0-d35b-4259-aebf-ab7b67e15ff2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"78be5076acc5fd27bf1c8ccfa9b3bf55b66b0d147ea7730764d8cdbee7b95868\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-5gp9s" podUID="d13c79a0-d35b-4259-aebf-ab7b67e15ff2" May 15 15:13:51.722368 kubelet[2694]: E0515 15:13:51.721326 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:51.724389 containerd[1549]: time="2025-05-15T15:13:51.724224868Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" May 15 15:13:52.491393 systemd[1]: Created slice kubepods-besteffort-podea7342ed_f869_4f54_b67c_23b40f5f8961.slice - libcontainer container kubepods-besteffort-podea7342ed_f869_4f54_b67c_23b40f5f8961.slice. May 15 15:13:52.496431 containerd[1549]: time="2025-05-15T15:13:52.496192931Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5mw49,Uid:ea7342ed-f869-4f54-b67c-23b40f5f8961,Namespace:calico-system,Attempt:0,}" May 15 15:13:52.552946 containerd[1549]: time="2025-05-15T15:13:52.552851022Z" level=error msg="Failed to destroy network for sandbox \"18478e331344fd99e971ff44cf926a0cc4258e7ec1219d6868e2b0190079e530\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:52.555751 containerd[1549]: time="2025-05-15T15:13:52.555689315Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5mw49,Uid:ea7342ed-f869-4f54-b67c-23b40f5f8961,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"18478e331344fd99e971ff44cf926a0cc4258e7ec1219d6868e2b0190079e530\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:52.556154 kubelet[2694]: E0515 15:13:52.556112 2694 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"18478e331344fd99e971ff44cf926a0cc4258e7ec1219d6868e2b0190079e530\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 15:13:52.556816 kubelet[2694]: E0515 15:13:52.556484 2694 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"18478e331344fd99e971ff44cf926a0cc4258e7ec1219d6868e2b0190079e530\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5mw49" May 15 15:13:52.556816 kubelet[2694]: E0515 15:13:52.556514 2694 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"18478e331344fd99e971ff44cf926a0cc4258e7ec1219d6868e2b0190079e530\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5mw49" May 15 15:13:52.556816 kubelet[2694]: E0515 15:13:52.556561 2694 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-5mw49_calico-system(ea7342ed-f869-4f54-b67c-23b40f5f8961)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-5mw49_calico-system(ea7342ed-f869-4f54-b67c-23b40f5f8961)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"18478e331344fd99e971ff44cf926a0cc4258e7ec1219d6868e2b0190079e530\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-5mw49" podUID="ea7342ed-f869-4f54-b67c-23b40f5f8961" May 15 15:13:52.556605 systemd[1]: run-netns-cni\x2d53ce68da\x2d25f0\x2d5c01\x2d302d\x2d692e6c671697.mount: Deactivated successfully. May 15 15:13:58.836153 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount461765217.mount: Deactivated successfully. May 15 15:13:58.868255 containerd[1549]: time="2025-05-15T15:13:58.868051168Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:58.869046 containerd[1549]: time="2025-05-15T15:13:58.869011003Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.3: active requests=0, bytes read=144068748" May 15 15:13:58.869659 containerd[1549]: time="2025-05-15T15:13:58.869628543Z" level=info msg="ImageCreate event name:\"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:58.871132 containerd[1549]: time="2025-05-15T15:13:58.871102604Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:13:58.871651 containerd[1549]: time="2025-05-15T15:13:58.871625844Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.3\" with image id \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\", size \"144068610\" in 7.147362233s" May 15 15:13:58.871700 containerd[1549]: time="2025-05-15T15:13:58.871656182Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\"" May 15 15:13:58.898331 containerd[1549]: time="2025-05-15T15:13:58.898293926Z" level=info msg="CreateContainer within sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 15 15:13:58.907233 containerd[1549]: time="2025-05-15T15:13:58.907183289Z" level=info msg="Container c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e: CDI devices from CRI Config.CDIDevices: []" May 15 15:13:58.915894 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3177551647.mount: Deactivated successfully. May 15 15:13:58.920122 containerd[1549]: time="2025-05-15T15:13:58.919971615Z" level=info msg="CreateContainer within sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\"" May 15 15:13:58.921304 containerd[1549]: time="2025-05-15T15:13:58.920886800Z" level=info msg="StartContainer for \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\"" May 15 15:13:58.923285 containerd[1549]: time="2025-05-15T15:13:58.923252848Z" level=info msg="connecting to shim c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e" address="unix:///run/containerd/s/bc46e7a4428a679462b18c2d1cb7901b190dc09c18f931c75b7f5148f6288437" protocol=ttrpc version=3 May 15 15:13:58.995177 systemd[1]: Started cri-containerd-c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e.scope - libcontainer container c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e. May 15 15:13:59.047475 containerd[1549]: time="2025-05-15T15:13:59.047414593Z" level=info msg="StartContainer for \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" returns successfully" May 15 15:13:59.290973 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. May 15 15:13:59.291136 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. May 15 15:13:59.755849 kubelet[2694]: E0515 15:13:59.755751 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:13:59.777300 kubelet[2694]: I0515 15:13:59.777146 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-qjmm6" podStartSLOduration=1.776462073 podStartE2EDuration="21.776995501s" podCreationTimestamp="2025-05-15 15:13:38 +0000 UTC" firstStartedPulling="2025-05-15 15:13:38.87988267 +0000 UTC m=+12.560755676" lastFinishedPulling="2025-05-15 15:13:58.880416096 +0000 UTC m=+32.561289104" observedRunningTime="2025-05-15 15:13:59.775461553 +0000 UTC m=+33.456334580" watchObservedRunningTime="2025-05-15 15:13:59.776995501 +0000 UTC m=+33.457868529" May 15 15:14:00.758473 kubelet[2694]: I0515 15:14:00.758427 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:00.760598 kubelet[2694]: E0515 15:14:00.760192 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:01.398139 containerd[1549]: time="2025-05-15T15:14:01.397316351Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" id:\"de9a94ce625e6a4f3b6addea39a60b775d6eff9b1ae27760892f424200b0a4d5\" pid:3861 exit_status:1 exited_at:{seconds:1747322041 nanos:386387311}" May 15 15:14:01.534236 containerd[1549]: time="2025-05-15T15:14:01.534188188Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" id:\"057bb08649264a53aed439b265774cd27d014d80e4a71ae1ecff1903335e627c\" pid:3884 exit_status:1 exited_at:{seconds:1747322041 nanos:533750582}" May 15 15:14:01.767917 kubelet[2694]: E0515 15:14:01.767877 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:01.843600 containerd[1549]: time="2025-05-15T15:14:01.843518845Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" id:\"b45f5f0f35701fdc8fb42bd2503db9bfd4ebc1cc22aaeff65a6d17f013d2986a\" pid:3910 exit_status:1 exited_at:{seconds:1747322041 nanos:842765389}" May 15 15:14:02.482448 kubelet[2694]: E0515 15:14:02.482336 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:02.483719 containerd[1549]: time="2025-05-15T15:14:02.483667535Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-c8cd4b68c-85hmc,Uid:826ead8b-3406-4df0-8588-b2656fcca004,Namespace:calico-system,Attempt:0,}" May 15 15:14:02.485127 containerd[1549]: time="2025-05-15T15:14:02.484629753Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vrxcl,Uid:64abed57-9eb1-4790-953d-d2c240a13afe,Namespace:kube-system,Attempt:0,}" May 15 15:14:02.814239 systemd-networkd[1455]: calidacd95855ff: Link UP May 15 15:14:02.814871 systemd-networkd[1455]: calidacd95855ff: Gained carrier May 15 15:14:02.835755 containerd[1549]: 2025-05-15 15:14:02.562 [INFO][3942] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 15 15:14:02.835755 containerd[1549]: 2025-05-15 15:14:02.591 [INFO][3942] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0 calico-kube-controllers-c8cd4b68c- calico-system 826ead8b-3406-4df0-8588-b2656fcca004 745 0 2025-05-15 15:13:38 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:c8cd4b68c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a calico-kube-controllers-c8cd4b68c-85hmc eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calidacd95855ff [] []}} ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Namespace="calico-system" Pod="calico-kube-controllers-c8cd4b68c-85hmc" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-" May 15 15:14:02.835755 containerd[1549]: 2025-05-15 15:14:02.591 [INFO][3942] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Namespace="calico-system" Pod="calico-kube-controllers-c8cd4b68c-85hmc" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:02.835755 containerd[1549]: 2025-05-15 15:14:02.732 [INFO][3967] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" HandleID="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.754 [INFO][3967] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" HandleID="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000398750), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4334.0.0-a-073c24674a", "pod":"calico-kube-controllers-c8cd4b68c-85hmc", "timestamp":"2025-05-15 15:14:02.73228116 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.754 [INFO][3967] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.755 [INFO][3967] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.755 [INFO][3967] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.758 [INFO][3967] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.767 [INFO][3967] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.774 [INFO][3967] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.776 [INFO][3967] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836253 containerd[1549]: 2025-05-15 15:14:02.779 [INFO][3967] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836502 containerd[1549]: 2025-05-15 15:14:02.779 [INFO][3967] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836502 containerd[1549]: 2025-05-15 15:14:02.782 [INFO][3967] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f May 15 15:14:02.836502 containerd[1549]: 2025-05-15 15:14:02.787 [INFO][3967] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836502 containerd[1549]: 2025-05-15 15:14:02.795 [INFO][3967] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.65/26] block=192.168.50.64/26 handle="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836502 containerd[1549]: 2025-05-15 15:14:02.795 [INFO][3967] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.65/26] handle="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.836502 containerd[1549]: 2025-05-15 15:14:02.795 [INFO][3967] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:02.836502 containerd[1549]: 2025-05-15 15:14:02.795 [INFO][3967] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.65/26] IPv6=[] ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" HandleID="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:02.840125 containerd[1549]: 2025-05-15 15:14:02.799 [INFO][3942] cni-plugin/k8s.go 386: Populated endpoint ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Namespace="calico-system" Pod="calico-kube-controllers-c8cd4b68c-85hmc" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0", GenerateName:"calico-kube-controllers-c8cd4b68c-", Namespace:"calico-system", SelfLink:"", UID:"826ead8b-3406-4df0-8588-b2656fcca004", ResourceVersion:"745", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"c8cd4b68c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"calico-kube-controllers-c8cd4b68c-85hmc", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.50.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calidacd95855ff", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:02.840216 containerd[1549]: 2025-05-15 15:14:02.799 [INFO][3942] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.65/32] ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Namespace="calico-system" Pod="calico-kube-controllers-c8cd4b68c-85hmc" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:02.840216 containerd[1549]: 2025-05-15 15:14:02.799 [INFO][3942] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidacd95855ff ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Namespace="calico-system" Pod="calico-kube-controllers-c8cd4b68c-85hmc" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:02.840216 containerd[1549]: 2025-05-15 15:14:02.817 [INFO][3942] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Namespace="calico-system" Pod="calico-kube-controllers-c8cd4b68c-85hmc" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:02.840289 containerd[1549]: 2025-05-15 15:14:02.818 [INFO][3942] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Namespace="calico-system" Pod="calico-kube-controllers-c8cd4b68c-85hmc" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0", GenerateName:"calico-kube-controllers-c8cd4b68c-", Namespace:"calico-system", SelfLink:"", UID:"826ead8b-3406-4df0-8588-b2656fcca004", ResourceVersion:"745", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"c8cd4b68c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f", Pod:"calico-kube-controllers-c8cd4b68c-85hmc", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.50.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calidacd95855ff", MAC:"9a:b0:33:96:42:73", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:02.842058 containerd[1549]: 2025-05-15 15:14:02.829 [INFO][3942] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Namespace="calico-system" Pod="calico-kube-controllers-c8cd4b68c-85hmc" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:02.907361 systemd-networkd[1455]: caliad4d86af1b3: Link UP May 15 15:14:02.908228 systemd-networkd[1455]: caliad4d86af1b3: Gained carrier May 15 15:14:02.931193 containerd[1549]: time="2025-05-15T15:14:02.931150282Z" level=info msg="connecting to shim a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" address="unix:///run/containerd/s/d4e122f5db653f848faceabcabd270c13e1a37235e1a190a20c59a5b92d3e48b" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:02.942278 containerd[1549]: 2025-05-15 15:14:02.564 [INFO][3943] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 15 15:14:02.942278 containerd[1549]: 2025-05-15 15:14:02.596 [INFO][3943] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0 coredns-668d6bf9bc- kube-system 64abed57-9eb1-4790-953d-d2c240a13afe 739 0 2025-05-15 15:13:31 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a coredns-668d6bf9bc-vrxcl eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] caliad4d86af1b3 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Namespace="kube-system" Pod="coredns-668d6bf9bc-vrxcl" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-" May 15 15:14:02.942278 containerd[1549]: 2025-05-15 15:14:02.596 [INFO][3943] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Namespace="kube-system" Pod="coredns-668d6bf9bc-vrxcl" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" May 15 15:14:02.942278 containerd[1549]: 2025-05-15 15:14:02.732 [INFO][3965] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" HandleID="k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Workload="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.755 [INFO][3965] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" HandleID="k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Workload="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003a5560), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4334.0.0-a-073c24674a", "pod":"coredns-668d6bf9bc-vrxcl", "timestamp":"2025-05-15 15:14:02.732385938 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.755 [INFO][3965] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.795 [INFO][3965] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.795 [INFO][3965] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.859 [INFO][3965] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.867 [INFO][3965] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.876 [INFO][3965] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.879 [INFO][3965] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.942676 containerd[1549]: 2025-05-15 15:14:02.882 [INFO][3965] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.943620 containerd[1549]: 2025-05-15 15:14:02.882 [INFO][3965] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.943620 containerd[1549]: 2025-05-15 15:14:02.884 [INFO][3965] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08 May 15 15:14:02.943620 containerd[1549]: 2025-05-15 15:14:02.890 [INFO][3965] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.943620 containerd[1549]: 2025-05-15 15:14:02.896 [INFO][3965] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.66/26] block=192.168.50.64/26 handle="k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.943620 containerd[1549]: 2025-05-15 15:14:02.897 [INFO][3965] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.66/26] handle="k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" host="ci-4334.0.0-a-073c24674a" May 15 15:14:02.943620 containerd[1549]: 2025-05-15 15:14:02.898 [INFO][3965] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:02.943620 containerd[1549]: 2025-05-15 15:14:02.898 [INFO][3965] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.66/26] IPv6=[] ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" HandleID="k8s-pod-network.560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Workload="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" May 15 15:14:02.943902 containerd[1549]: 2025-05-15 15:14:02.903 [INFO][3943] cni-plugin/k8s.go 386: Populated endpoint ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Namespace="kube-system" Pod="coredns-668d6bf9bc-vrxcl" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"64abed57-9eb1-4790-953d-d2c240a13afe", ResourceVersion:"739", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"coredns-668d6bf9bc-vrxcl", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.50.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliad4d86af1b3", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:02.943902 containerd[1549]: 2025-05-15 15:14:02.903 [INFO][3943] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.66/32] ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Namespace="kube-system" Pod="coredns-668d6bf9bc-vrxcl" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" May 15 15:14:02.943902 containerd[1549]: 2025-05-15 15:14:02.903 [INFO][3943] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliad4d86af1b3 ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Namespace="kube-system" Pod="coredns-668d6bf9bc-vrxcl" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" May 15 15:14:02.943902 containerd[1549]: 2025-05-15 15:14:02.908 [INFO][3943] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Namespace="kube-system" Pod="coredns-668d6bf9bc-vrxcl" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" May 15 15:14:02.943902 containerd[1549]: 2025-05-15 15:14:02.909 [INFO][3943] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Namespace="kube-system" Pod="coredns-668d6bf9bc-vrxcl" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"64abed57-9eb1-4790-953d-d2c240a13afe", ResourceVersion:"739", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08", Pod:"coredns-668d6bf9bc-vrxcl", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.50.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliad4d86af1b3", MAC:"56:57:2e:fe:d7:a6", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:02.943902 containerd[1549]: 2025-05-15 15:14:02.933 [INFO][3943] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" Namespace="kube-system" Pod="coredns-668d6bf9bc-vrxcl" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--vrxcl-eth0" May 15 15:14:03.008115 containerd[1549]: time="2025-05-15T15:14:03.008042874Z" level=info msg="connecting to shim 560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08" address="unix:///run/containerd/s/893c5fa25a6bc86b444d46bff9b1b544d6db90d8c8638431b0447b9e561504cc" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:03.023471 systemd[1]: Started cri-containerd-a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f.scope - libcontainer container a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f. May 15 15:14:03.069713 systemd[1]: Started cri-containerd-560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08.scope - libcontainer container 560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08. May 15 15:14:03.135744 containerd[1549]: time="2025-05-15T15:14:03.135701434Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vrxcl,Uid:64abed57-9eb1-4790-953d-d2c240a13afe,Namespace:kube-system,Attempt:0,} returns sandbox id \"560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08\"" May 15 15:14:03.136773 kubelet[2694]: E0515 15:14:03.136747 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:03.146179 containerd[1549]: time="2025-05-15T15:14:03.145544475Z" level=info msg="CreateContainer within sandbox \"560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 15 15:14:03.185320 containerd[1549]: time="2025-05-15T15:14:03.185278946Z" level=info msg="Container 48544307c0a6dea3d63cd0d5fb2a78d969dcae865816857af3ca3c600ad1bb2d: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:03.186838 containerd[1549]: time="2025-05-15T15:14:03.186802604Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-c8cd4b68c-85hmc,Uid:826ead8b-3406-4df0-8588-b2656fcca004,Namespace:calico-system,Attempt:0,} returns sandbox id \"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\"" May 15 15:14:03.190441 containerd[1549]: time="2025-05-15T15:14:03.190196029Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" May 15 15:14:03.197033 containerd[1549]: time="2025-05-15T15:14:03.195885161Z" level=info msg="CreateContainer within sandbox \"560577492e28c923e219e517d4561948e60ca1dd38b18d41cd268fd4dff86b08\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"48544307c0a6dea3d63cd0d5fb2a78d969dcae865816857af3ca3c600ad1bb2d\"" May 15 15:14:03.198553 containerd[1549]: time="2025-05-15T15:14:03.197461582Z" level=info msg="StartContainer for \"48544307c0a6dea3d63cd0d5fb2a78d969dcae865816857af3ca3c600ad1bb2d\"" May 15 15:14:03.200082 containerd[1549]: time="2025-05-15T15:14:03.199381987Z" level=info msg="connecting to shim 48544307c0a6dea3d63cd0d5fb2a78d969dcae865816857af3ca3c600ad1bb2d" address="unix:///run/containerd/s/893c5fa25a6bc86b444d46bff9b1b544d6db90d8c8638431b0447b9e561504cc" protocol=ttrpc version=3 May 15 15:14:03.233409 systemd[1]: Started cri-containerd-48544307c0a6dea3d63cd0d5fb2a78d969dcae865816857af3ca3c600ad1bb2d.scope - libcontainer container 48544307c0a6dea3d63cd0d5fb2a78d969dcae865816857af3ca3c600ad1bb2d. May 15 15:14:03.295815 containerd[1549]: time="2025-05-15T15:14:03.295762918Z" level=info msg="StartContainer for \"48544307c0a6dea3d63cd0d5fb2a78d969dcae865816857af3ca3c600ad1bb2d\" returns successfully" May 15 15:14:03.483651 containerd[1549]: time="2025-05-15T15:14:03.483604969Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6999544d4f-js9sf,Uid:869a7900-ca01-45de-9a8f-19a49b2d754b,Namespace:calico-apiserver,Attempt:0,}" May 15 15:14:03.718748 systemd-networkd[1455]: cali9e2522bce8a: Link UP May 15 15:14:03.719097 systemd-networkd[1455]: cali9e2522bce8a: Gained carrier May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.532 [INFO][4148] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.551 [INFO][4148] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0 calico-apiserver-6999544d4f- calico-apiserver 869a7900-ca01-45de-9a8f-19a49b2d754b 748 0 2025-05-15 15:13:38 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6999544d4f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a calico-apiserver-6999544d4f-js9sf eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali9e2522bce8a [] []}} ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-js9sf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.551 [INFO][4148] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-js9sf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.617 [INFO][4162] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" HandleID="k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.635 [INFO][4162] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" HandleID="k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003bc130), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4334.0.0-a-073c24674a", "pod":"calico-apiserver-6999544d4f-js9sf", "timestamp":"2025-05-15 15:14:03.617190873 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.635 [INFO][4162] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.635 [INFO][4162] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.636 [INFO][4162] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.650 [INFO][4162] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.657 [INFO][4162] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.666 [INFO][4162] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.673 [INFO][4162] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.678 [INFO][4162] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.679 [INFO][4162] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.681 [INFO][4162] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.689 [INFO][4162] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.704 [INFO][4162] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.67/26] block=192.168.50.64/26 handle="k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.704 [INFO][4162] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.67/26] handle="k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" host="ci-4334.0.0-a-073c24674a" May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.704 [INFO][4162] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:03.742498 containerd[1549]: 2025-05-15 15:14:03.704 [INFO][4162] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.67/26] IPv6=[] ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" HandleID="k8s-pod-network.bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" May 15 15:14:03.745488 containerd[1549]: 2025-05-15 15:14:03.710 [INFO][4148] cni-plugin/k8s.go 386: Populated endpoint ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-js9sf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0", GenerateName:"calico-apiserver-6999544d4f-", Namespace:"calico-apiserver", SelfLink:"", UID:"869a7900-ca01-45de-9a8f-19a49b2d754b", ResourceVersion:"748", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6999544d4f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"calico-apiserver-6999544d4f-js9sf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.50.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9e2522bce8a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:03.745488 containerd[1549]: 2025-05-15 15:14:03.710 [INFO][4148] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.67/32] ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-js9sf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" May 15 15:14:03.745488 containerd[1549]: 2025-05-15 15:14:03.710 [INFO][4148] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9e2522bce8a ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-js9sf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" May 15 15:14:03.745488 containerd[1549]: 2025-05-15 15:14:03.715 [INFO][4148] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-js9sf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" May 15 15:14:03.745488 containerd[1549]: 2025-05-15 15:14:03.718 [INFO][4148] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-js9sf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0", GenerateName:"calico-apiserver-6999544d4f-", Namespace:"calico-apiserver", SelfLink:"", UID:"869a7900-ca01-45de-9a8f-19a49b2d754b", ResourceVersion:"748", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6999544d4f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa", Pod:"calico-apiserver-6999544d4f-js9sf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.50.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9e2522bce8a", MAC:"de:38:e5:3a:34:bd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:03.745488 containerd[1549]: 2025-05-15 15:14:03.734 [INFO][4148] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-js9sf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--js9sf-eth0" May 15 15:14:03.781469 kubelet[2694]: E0515 15:14:03.779770 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:03.794237 containerd[1549]: time="2025-05-15T15:14:03.793826051Z" level=info msg="connecting to shim bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa" address="unix:///run/containerd/s/a1067ee71aa73c5fc7a225907f7026c713633554235ddb9ecb04ac3184fd5cf6" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:03.821676 kubelet[2694]: I0515 15:14:03.821304 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-vrxcl" podStartSLOduration=32.821281264 podStartE2EDuration="32.821281264s" podCreationTimestamp="2025-05-15 15:13:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:14:03.820601252 +0000 UTC m=+37.501474279" watchObservedRunningTime="2025-05-15 15:14:03.821281264 +0000 UTC m=+37.502154285" May 15 15:14:03.841147 kubelet[2694]: I0515 15:14:03.841005 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:03.844233 kubelet[2694]: E0515 15:14:03.844195 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:03.866272 systemd[1]: Started cri-containerd-bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa.scope - libcontainer container bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa. May 15 15:14:04.017492 containerd[1549]: time="2025-05-15T15:14:04.017240682Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6999544d4f-js9sf,Uid:869a7900-ca01-45de-9a8f-19a49b2d754b,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa\"" May 15 15:14:04.239616 systemd-networkd[1455]: calidacd95855ff: Gained IPv6LL May 15 15:14:04.303475 systemd-networkd[1455]: caliad4d86af1b3: Gained IPv6LL May 15 15:14:04.483002 kubelet[2694]: E0515 15:14:04.482938 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:04.485402 containerd[1549]: time="2025-05-15T15:14:04.485361328Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5gp9s,Uid:d13c79a0-d35b-4259-aebf-ab7b67e15ff2,Namespace:kube-system,Attempt:0,}" May 15 15:14:04.648292 systemd-networkd[1455]: calif3bb23facdf: Link UP May 15 15:14:04.648921 systemd-networkd[1455]: calif3bb23facdf: Gained carrier May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.549 [INFO][4278] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0 coredns-668d6bf9bc- kube-system d13c79a0-d35b-4259-aebf-ab7b67e15ff2 746 0 2025-05-15 15:13:31 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a coredns-668d6bf9bc-5gp9s eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calif3bb23facdf [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Namespace="kube-system" Pod="coredns-668d6bf9bc-5gp9s" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.549 [INFO][4278] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Namespace="kube-system" Pod="coredns-668d6bf9bc-5gp9s" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.589 [INFO][4298] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" HandleID="k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Workload="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.601 [INFO][4298] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" HandleID="k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Workload="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0004252c0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4334.0.0-a-073c24674a", "pod":"coredns-668d6bf9bc-5gp9s", "timestamp":"2025-05-15 15:14:04.589555932 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.601 [INFO][4298] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.601 [INFO][4298] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.601 [INFO][4298] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.604 [INFO][4298] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.610 [INFO][4298] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.616 [INFO][4298] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.619 [INFO][4298] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.623 [INFO][4298] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.623 [INFO][4298] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.625 [INFO][4298] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414 May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.631 [INFO][4298] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.640 [INFO][4298] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.68/26] block=192.168.50.64/26 handle="k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.641 [INFO][4298] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.68/26] handle="k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" host="ci-4334.0.0-a-073c24674a" May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.641 [INFO][4298] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:04.670286 containerd[1549]: 2025-05-15 15:14:04.641 [INFO][4298] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.68/26] IPv6=[] ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" HandleID="k8s-pod-network.50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Workload="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" May 15 15:14:04.672703 containerd[1549]: 2025-05-15 15:14:04.644 [INFO][4278] cni-plugin/k8s.go 386: Populated endpoint ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Namespace="kube-system" Pod="coredns-668d6bf9bc-5gp9s" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"d13c79a0-d35b-4259-aebf-ab7b67e15ff2", ResourceVersion:"746", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"coredns-668d6bf9bc-5gp9s", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.50.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif3bb23facdf", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:04.672703 containerd[1549]: 2025-05-15 15:14:04.644 [INFO][4278] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.68/32] ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Namespace="kube-system" Pod="coredns-668d6bf9bc-5gp9s" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" May 15 15:14:04.672703 containerd[1549]: 2025-05-15 15:14:04.644 [INFO][4278] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif3bb23facdf ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Namespace="kube-system" Pod="coredns-668d6bf9bc-5gp9s" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" May 15 15:14:04.672703 containerd[1549]: 2025-05-15 15:14:04.649 [INFO][4278] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Namespace="kube-system" Pod="coredns-668d6bf9bc-5gp9s" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" May 15 15:14:04.672703 containerd[1549]: 2025-05-15 15:14:04.650 [INFO][4278] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Namespace="kube-system" Pod="coredns-668d6bf9bc-5gp9s" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"d13c79a0-d35b-4259-aebf-ab7b67e15ff2", ResourceVersion:"746", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414", Pod:"coredns-668d6bf9bc-5gp9s", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.50.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif3bb23facdf", MAC:"6a:e4:54:18:bf:55", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:04.672703 containerd[1549]: 2025-05-15 15:14:04.664 [INFO][4278] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" Namespace="kube-system" Pod="coredns-668d6bf9bc-5gp9s" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-coredns--668d6bf9bc--5gp9s-eth0" May 15 15:14:04.726891 containerd[1549]: time="2025-05-15T15:14:04.726381630Z" level=info msg="connecting to shim 50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414" address="unix:///run/containerd/s/51ba51ea995477563df029e77470e30fb4e02253f089de76fa42b4d068bde35e" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:04.769380 systemd[1]: Started cri-containerd-50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414.scope - libcontainer container 50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414. May 15 15:14:04.817546 kubelet[2694]: E0515 15:14:04.817377 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:04.818158 kubelet[2694]: E0515 15:14:04.818118 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:04.914388 systemd-networkd[1455]: vxlan.calico: Link UP May 15 15:14:04.914398 systemd-networkd[1455]: vxlan.calico: Gained carrier May 15 15:14:04.933458 containerd[1549]: time="2025-05-15T15:14:04.933313298Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5gp9s,Uid:d13c79a0-d35b-4259-aebf-ab7b67e15ff2,Namespace:kube-system,Attempt:0,} returns sandbox id \"50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414\"" May 15 15:14:04.937997 kubelet[2694]: E0515 15:14:04.935700 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:04.939624 containerd[1549]: time="2025-05-15T15:14:04.939416012Z" level=info msg="CreateContainer within sandbox \"50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 15 15:14:04.967258 containerd[1549]: time="2025-05-15T15:14:04.966685070Z" level=info msg="Container 08b59d292340ec8d9d2ae1c6c946a6b75604ad7451f054c65d0012a06cdc1f1f: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:04.991840 containerd[1549]: time="2025-05-15T15:14:04.991786980Z" level=info msg="CreateContainer within sandbox \"50756119894c768f0caef898efd46bc1fe4cc5a70f4f2d61d531de5026f17414\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"08b59d292340ec8d9d2ae1c6c946a6b75604ad7451f054c65d0012a06cdc1f1f\"" May 15 15:14:04.994779 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3674367352.mount: Deactivated successfully. May 15 15:14:04.998880 containerd[1549]: time="2025-05-15T15:14:04.998828590Z" level=info msg="StartContainer for \"08b59d292340ec8d9d2ae1c6c946a6b75604ad7451f054c65d0012a06cdc1f1f\"" May 15 15:14:05.005396 containerd[1549]: time="2025-05-15T15:14:05.002557888Z" level=info msg="connecting to shim 08b59d292340ec8d9d2ae1c6c946a6b75604ad7451f054c65d0012a06cdc1f1f" address="unix:///run/containerd/s/51ba51ea995477563df029e77470e30fb4e02253f089de76fa42b4d068bde35e" protocol=ttrpc version=3 May 15 15:14:05.073375 systemd[1]: Started cri-containerd-08b59d292340ec8d9d2ae1c6c946a6b75604ad7451f054c65d0012a06cdc1f1f.scope - libcontainer container 08b59d292340ec8d9d2ae1c6c946a6b75604ad7451f054c65d0012a06cdc1f1f. May 15 15:14:05.131021 containerd[1549]: time="2025-05-15T15:14:05.130984407Z" level=info msg="StartContainer for \"08b59d292340ec8d9d2ae1c6c946a6b75604ad7451f054c65d0012a06cdc1f1f\" returns successfully" May 15 15:14:05.483909 containerd[1549]: time="2025-05-15T15:14:05.483853296Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-585b9b4df4-cr2q4,Uid:52c8dc92-1a03-4e1d-bc5b-3844d9a477ad,Namespace:calico-apiserver,Attempt:0,}" May 15 15:14:05.519142 systemd-networkd[1455]: cali9e2522bce8a: Gained IPv6LL May 15 15:14:05.765387 systemd-networkd[1455]: cali743a39b7768: Link UP May 15 15:14:05.768144 systemd-networkd[1455]: cali743a39b7768: Gained carrier May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.572 [INFO][4442] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0 calico-apiserver-585b9b4df4- calico-apiserver 52c8dc92-1a03-4e1d-bc5b-3844d9a477ad 747 0 2025-05-15 15:13:38 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:585b9b4df4 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a calico-apiserver-585b9b4df4-cr2q4 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali743a39b7768 [] []}} ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-cr2q4" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.572 [INFO][4442] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-cr2q4" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.676 [INFO][4472] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.693 [INFO][4472] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000374e30), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4334.0.0-a-073c24674a", "pod":"calico-apiserver-585b9b4df4-cr2q4", "timestamp":"2025-05-15 15:14:05.676063042 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.694 [INFO][4472] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.694 [INFO][4472] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.694 [INFO][4472] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.699 [INFO][4472] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.709 [INFO][4472] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.720 [INFO][4472] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.724 [INFO][4472] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.729 [INFO][4472] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.729 [INFO][4472] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.732 [INFO][4472] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611 May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.740 [INFO][4472] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.749 [INFO][4472] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.69/26] block=192.168.50.64/26 handle="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.750 [INFO][4472] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.69/26] handle="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" host="ci-4334.0.0-a-073c24674a" May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.750 [INFO][4472] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:05.806210 containerd[1549]: 2025-05-15 15:14:05.750 [INFO][4472] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.69/26] IPv6=[] ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:05.808640 containerd[1549]: 2025-05-15 15:14:05.759 [INFO][4442] cni-plugin/k8s.go 386: Populated endpoint ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-cr2q4" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0", GenerateName:"calico-apiserver-585b9b4df4-", Namespace:"calico-apiserver", SelfLink:"", UID:"52c8dc92-1a03-4e1d-bc5b-3844d9a477ad", ResourceVersion:"747", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"585b9b4df4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"calico-apiserver-585b9b4df4-cr2q4", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.50.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali743a39b7768", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:05.808640 containerd[1549]: 2025-05-15 15:14:05.759 [INFO][4442] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.69/32] ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-cr2q4" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:05.808640 containerd[1549]: 2025-05-15 15:14:05.759 [INFO][4442] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali743a39b7768 ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-cr2q4" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:05.808640 containerd[1549]: 2025-05-15 15:14:05.769 [INFO][4442] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-cr2q4" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:05.808640 containerd[1549]: 2025-05-15 15:14:05.772 [INFO][4442] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-cr2q4" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0", GenerateName:"calico-apiserver-585b9b4df4-", Namespace:"calico-apiserver", SelfLink:"", UID:"52c8dc92-1a03-4e1d-bc5b-3844d9a477ad", ResourceVersion:"747", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"585b9b4df4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611", Pod:"calico-apiserver-585b9b4df4-cr2q4", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.50.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali743a39b7768", MAC:"b2:2c:71:2f:69:01", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:05.808640 containerd[1549]: 2025-05-15 15:14:05.799 [INFO][4442] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-cr2q4" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:05.826485 kubelet[2694]: E0515 15:14:05.826443 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:05.828621 kubelet[2694]: E0515 15:14:05.828592 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:05.866921 kubelet[2694]: I0515 15:14:05.866848 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-5gp9s" podStartSLOduration=34.866559825 podStartE2EDuration="34.866559825s" podCreationTimestamp="2025-05-15 15:13:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:14:05.86108494 +0000 UTC m=+39.541957963" watchObservedRunningTime="2025-05-15 15:14:05.866559825 +0000 UTC m=+39.547432851" May 15 15:14:05.885003 containerd[1549]: time="2025-05-15T15:14:05.884166239Z" level=info msg="connecting to shim 41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" address="unix:///run/containerd/s/3b9a209427bbd1d72961ddd54be4a7907f6bcc8df1f819ad250b54740797cb46" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:05.987339 systemd[1]: Started cri-containerd-41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611.scope - libcontainer container 41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611. May 15 15:14:06.075399 containerd[1549]: time="2025-05-15T15:14:06.074437053Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-585b9b4df4-cr2q4,Uid:52c8dc92-1a03-4e1d-bc5b-3844d9a477ad,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\"" May 15 15:14:06.223158 systemd-networkd[1455]: vxlan.calico: Gained IPv6LL May 15 15:14:06.412023 containerd[1549]: time="2025-05-15T15:14:06.411060634Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.3: active requests=0, bytes read=34789138" May 15 15:14:06.429386 containerd[1549]: time="2025-05-15T15:14:06.428285015Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" with image id \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\", size \"36281728\" in 3.228575413s" May 15 15:14:06.429386 containerd[1549]: time="2025-05-15T15:14:06.429286948Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\"" May 15 15:14:06.438286 containerd[1549]: time="2025-05-15T15:14:06.437382842Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 15 15:14:06.475211 containerd[1549]: time="2025-05-15T15:14:06.474531706Z" level=info msg="CreateContainer within sandbox \"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" May 15 15:14:06.481114 systemd-networkd[1455]: calif3bb23facdf: Gained IPv6LL May 15 15:14:06.490863 containerd[1549]: time="2025-05-15T15:14:06.490823168Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:06.498053 containerd[1549]: time="2025-05-15T15:14:06.498011332Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-585b9b4df4-d2bqf,Uid:1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7,Namespace:calico-apiserver,Attempt:0,}" May 15 15:14:06.502544 containerd[1549]: time="2025-05-15T15:14:06.502503168Z" level=info msg="ImageCreate event name:\"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:06.503016 containerd[1549]: time="2025-05-15T15:14:06.502993144Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:06.520996 containerd[1549]: time="2025-05-15T15:14:06.520250730Z" level=info msg="Container 6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:06.564746 containerd[1549]: time="2025-05-15T15:14:06.564610201Z" level=info msg="CreateContainer within sandbox \"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\"" May 15 15:14:06.569137 containerd[1549]: time="2025-05-15T15:14:06.568761574Z" level=info msg="StartContainer for \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\"" May 15 15:14:06.584532 containerd[1549]: time="2025-05-15T15:14:06.576886273Z" level=info msg="connecting to shim 6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11" address="unix:///run/containerd/s/d4e122f5db653f848faceabcabd270c13e1a37235e1a190a20c59a5b92d3e48b" protocol=ttrpc version=3 May 15 15:14:06.647199 systemd[1]: Started cri-containerd-6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11.scope - libcontainer container 6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11. May 15 15:14:06.797920 systemd-networkd[1455]: caliabce917082c: Link UP May 15 15:14:06.801131 systemd-networkd[1455]: caliabce917082c: Gained carrier May 15 15:14:06.828766 containerd[1549]: time="2025-05-15T15:14:06.828141851Z" level=info msg="StartContainer for \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" returns successfully" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.645 [INFO][4557] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0 calico-apiserver-585b9b4df4- calico-apiserver 1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7 749 0 2025-05-15 15:13:38 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:585b9b4df4 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a calico-apiserver-585b9b4df4-d2bqf eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] caliabce917082c [] []}} ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-d2bqf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.645 [INFO][4557] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-d2bqf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.706 [INFO][4590] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.721 [INFO][4590] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000291230), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4334.0.0-a-073c24674a", "pod":"calico-apiserver-585b9b4df4-d2bqf", "timestamp":"2025-05-15 15:14:06.70669808 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.721 [INFO][4590] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.721 [INFO][4590] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.721 [INFO][4590] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.726 [INFO][4590] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.731 [INFO][4590] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.736 [INFO][4590] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.739 [INFO][4590] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.747 [INFO][4590] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.747 [INFO][4590] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.750 [INFO][4590] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30 May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.763 [INFO][4590] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.784 [INFO][4590] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.70/26] block=192.168.50.64/26 handle="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.784 [INFO][4590] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.70/26] handle="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" host="ci-4334.0.0-a-073c24674a" May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.784 [INFO][4590] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:06.841353 containerd[1549]: 2025-05-15 15:14:06.784 [INFO][4590] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.70/26] IPv6=[] ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:06.844334 containerd[1549]: 2025-05-15 15:14:06.789 [INFO][4557] cni-plugin/k8s.go 386: Populated endpoint ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-d2bqf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0", GenerateName:"calico-apiserver-585b9b4df4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7", ResourceVersion:"749", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"585b9b4df4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"calico-apiserver-585b9b4df4-d2bqf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.50.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliabce917082c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:06.844334 containerd[1549]: 2025-05-15 15:14:06.790 [INFO][4557] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.70/32] ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-d2bqf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:06.844334 containerd[1549]: 2025-05-15 15:14:06.790 [INFO][4557] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliabce917082c ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-d2bqf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:06.844334 containerd[1549]: 2025-05-15 15:14:06.802 [INFO][4557] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-d2bqf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:06.844334 containerd[1549]: 2025-05-15 15:14:06.804 [INFO][4557] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-d2bqf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0", GenerateName:"calico-apiserver-585b9b4df4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7", ResourceVersion:"749", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"585b9b4df4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30", Pod:"calico-apiserver-585b9b4df4-d2bqf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.50.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliabce917082c", MAC:"da:9d:e9:89:8a:c7", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:06.844334 containerd[1549]: 2025-05-15 15:14:06.828 [INFO][4557] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Namespace="calico-apiserver" Pod="calico-apiserver-585b9b4df4-d2bqf" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:06.871712 kubelet[2694]: E0515 15:14:06.870659 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:06.911088 kubelet[2694]: I0515 15:14:06.910260 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-c8cd4b68c-85hmc" podStartSLOduration=25.665062548999998 podStartE2EDuration="28.910238705s" podCreationTimestamp="2025-05-15 15:13:38 +0000 UTC" firstStartedPulling="2025-05-15 15:14:03.189365064 +0000 UTC m=+36.870238070" lastFinishedPulling="2025-05-15 15:14:06.434541212 +0000 UTC m=+40.115414226" observedRunningTime="2025-05-15 15:14:06.908923299 +0000 UTC m=+40.589796326" watchObservedRunningTime="2025-05-15 15:14:06.910238705 +0000 UTC m=+40.591111732" May 15 15:14:06.946942 containerd[1549]: time="2025-05-15T15:14:06.946848799Z" level=info msg="connecting to shim 5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" address="unix:///run/containerd/s/4abf4ae9599c0886af21aef382ac4f41aee7e4460afebcdffc2ce2a9f0037e0a" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:07.008202 systemd[1]: Started cri-containerd-5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30.scope - libcontainer container 5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30. May 15 15:14:07.100072 containerd[1549]: time="2025-05-15T15:14:07.100021616Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-585b9b4df4-d2bqf,Uid:1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\"" May 15 15:14:07.311169 systemd-networkd[1455]: cali743a39b7768: Gained IPv6LL May 15 15:14:07.483090 containerd[1549]: time="2025-05-15T15:14:07.482984113Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5mw49,Uid:ea7342ed-f869-4f54-b67c-23b40f5f8961,Namespace:calico-system,Attempt:0,}" May 15 15:14:07.640546 systemd-networkd[1455]: cali2a83a5538a0: Link UP May 15 15:14:07.641666 systemd-networkd[1455]: cali2a83a5538a0: Gained carrier May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.532 [INFO][4667] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0 csi-node-driver- calico-system ea7342ed-f869-4f54-b67c-23b40f5f8961 626 0 2025-05-15 15:13:38 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:5b5cc68cd5 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a csi-node-driver-5mw49 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali2a83a5538a0 [] []}} ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Namespace="calico-system" Pod="csi-node-driver-5mw49" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.532 [INFO][4667] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Namespace="calico-system" Pod="csi-node-driver-5mw49" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.579 [INFO][4679] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" HandleID="k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Workload="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.590 [INFO][4679] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" HandleID="k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Workload="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000290e90), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4334.0.0-a-073c24674a", "pod":"csi-node-driver-5mw49", "timestamp":"2025-05-15 15:14:07.579053957 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.590 [INFO][4679] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.590 [INFO][4679] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.590 [INFO][4679] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.593 [INFO][4679] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.599 [INFO][4679] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.606 [INFO][4679] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.609 [INFO][4679] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.613 [INFO][4679] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.613 [INFO][4679] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.615 [INFO][4679] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409 May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.621 [INFO][4679] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.630 [INFO][4679] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.71/26] block=192.168.50.64/26 handle="k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.631 [INFO][4679] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.71/26] handle="k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" host="ci-4334.0.0-a-073c24674a" May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.631 [INFO][4679] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:07.670345 containerd[1549]: 2025-05-15 15:14:07.631 [INFO][4679] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.71/26] IPv6=[] ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" HandleID="k8s-pod-network.72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Workload="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" May 15 15:14:07.672517 containerd[1549]: 2025-05-15 15:14:07.635 [INFO][4667] cni-plugin/k8s.go 386: Populated endpoint ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Namespace="calico-system" Pod="csi-node-driver-5mw49" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ea7342ed-f869-4f54-b67c-23b40f5f8961", ResourceVersion:"626", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"csi-node-driver-5mw49", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.50.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2a83a5538a0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:07.672517 containerd[1549]: 2025-05-15 15:14:07.635 [INFO][4667] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.71/32] ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Namespace="calico-system" Pod="csi-node-driver-5mw49" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" May 15 15:14:07.672517 containerd[1549]: 2025-05-15 15:14:07.635 [INFO][4667] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2a83a5538a0 ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Namespace="calico-system" Pod="csi-node-driver-5mw49" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" May 15 15:14:07.672517 containerd[1549]: 2025-05-15 15:14:07.641 [INFO][4667] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Namespace="calico-system" Pod="csi-node-driver-5mw49" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" May 15 15:14:07.672517 containerd[1549]: 2025-05-15 15:14:07.641 [INFO][4667] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Namespace="calico-system" Pod="csi-node-driver-5mw49" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ea7342ed-f869-4f54-b67c-23b40f5f8961", ResourceVersion:"626", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 13, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409", Pod:"csi-node-driver-5mw49", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.50.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2a83a5538a0", MAC:"de:ac:c2:c2:64:53", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:07.672517 containerd[1549]: 2025-05-15 15:14:07.659 [INFO][4667] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" Namespace="calico-system" Pod="csi-node-driver-5mw49" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-csi--node--driver--5mw49-eth0" May 15 15:14:07.730712 containerd[1549]: time="2025-05-15T15:14:07.730620862Z" level=info msg="connecting to shim 72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409" address="unix:///run/containerd/s/121dcd7c2b44212bf606ebca34fa8945562113b9d906455d304b594015d3efe6" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:07.772266 systemd[1]: Started cri-containerd-72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409.scope - libcontainer container 72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409. May 15 15:14:07.823911 containerd[1549]: time="2025-05-15T15:14:07.823445471Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5mw49,Uid:ea7342ed-f869-4f54-b67c-23b40f5f8961,Namespace:calico-system,Attempt:0,} returns sandbox id \"72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409\"" May 15 15:14:07.874633 kubelet[2694]: I0515 15:14:07.874560 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:07.875252 kubelet[2694]: E0515 15:14:07.875063 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:08.271590 systemd-networkd[1455]: caliabce917082c: Gained IPv6LL May 15 15:14:09.295848 systemd-networkd[1455]: cali2a83a5538a0: Gained IPv6LL May 15 15:14:09.752751 containerd[1549]: time="2025-05-15T15:14:09.752663895Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:09.754392 containerd[1549]: time="2025-05-15T15:14:09.754352421Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=43021437" May 15 15:14:09.762270 containerd[1549]: time="2025-05-15T15:14:09.762168320Z" level=info msg="ImageCreate event name:\"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:09.764090 containerd[1549]: time="2025-05-15T15:14:09.764044583Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:09.765139 containerd[1549]: time="2025-05-15T15:14:09.765098961Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 3.326485937s" May 15 15:14:09.765139 containerd[1549]: time="2025-05-15T15:14:09.765138926Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 15 15:14:09.767587 containerd[1549]: time="2025-05-15T15:14:09.767555596Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 15 15:14:09.770593 containerd[1549]: time="2025-05-15T15:14:09.770553410Z" level=info msg="CreateContainer within sandbox \"bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 15 15:14:09.778954 containerd[1549]: time="2025-05-15T15:14:09.777556714Z" level=info msg="Container 1540e08a0bb3564068084cdd30e323e8fa0efce596c9f9024691b35a19c5d743: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:09.783500 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1824018854.mount: Deactivated successfully. May 15 15:14:09.809242 containerd[1549]: time="2025-05-15T15:14:09.809154727Z" level=info msg="CreateContainer within sandbox \"bc6b0eeacfdd3073e01dfaced7a72c8da1cc76ef6fd62a57c57d77941b6034fa\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"1540e08a0bb3564068084cdd30e323e8fa0efce596c9f9024691b35a19c5d743\"" May 15 15:14:09.810098 containerd[1549]: time="2025-05-15T15:14:09.809968223Z" level=info msg="StartContainer for \"1540e08a0bb3564068084cdd30e323e8fa0efce596c9f9024691b35a19c5d743\"" May 15 15:14:09.812265 containerd[1549]: time="2025-05-15T15:14:09.812183544Z" level=info msg="connecting to shim 1540e08a0bb3564068084cdd30e323e8fa0efce596c9f9024691b35a19c5d743" address="unix:///run/containerd/s/a1067ee71aa73c5fc7a225907f7026c713633554235ddb9ecb04ac3184fd5cf6" protocol=ttrpc version=3 May 15 15:14:09.846249 systemd[1]: Started cri-containerd-1540e08a0bb3564068084cdd30e323e8fa0efce596c9f9024691b35a19c5d743.scope - libcontainer container 1540e08a0bb3564068084cdd30e323e8fa0efce596c9f9024691b35a19c5d743. May 15 15:14:09.929162 containerd[1549]: time="2025-05-15T15:14:09.929054047Z" level=info msg="StartContainer for \"1540e08a0bb3564068084cdd30e323e8fa0efce596c9f9024691b35a19c5d743\" returns successfully" May 15 15:14:10.266370 containerd[1549]: time="2025-05-15T15:14:10.266298892Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:10.267059 containerd[1549]: time="2025-05-15T15:14:10.267005597Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=77" May 15 15:14:10.271725 containerd[1549]: time="2025-05-15T15:14:10.271535992Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 503.941792ms" May 15 15:14:10.271725 containerd[1549]: time="2025-05-15T15:14:10.271589857Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 15 15:14:10.273632 containerd[1549]: time="2025-05-15T15:14:10.273103023Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 15 15:14:10.274267 containerd[1549]: time="2025-05-15T15:14:10.274093589Z" level=info msg="CreateContainer within sandbox \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 15 15:14:10.288957 containerd[1549]: time="2025-05-15T15:14:10.288516052Z" level=info msg="Container 068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:10.292884 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4210869790.mount: Deactivated successfully. May 15 15:14:10.304915 containerd[1549]: time="2025-05-15T15:14:10.304842753Z" level=info msg="CreateContainer within sandbox \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\"" May 15 15:14:10.305797 containerd[1549]: time="2025-05-15T15:14:10.305717832Z" level=info msg="StartContainer for \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\"" May 15 15:14:10.307419 containerd[1549]: time="2025-05-15T15:14:10.307372257Z" level=info msg="connecting to shim 068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360" address="unix:///run/containerd/s/3b9a209427bbd1d72961ddd54be4a7907f6bcc8df1f819ad250b54740797cb46" protocol=ttrpc version=3 May 15 15:14:10.342268 systemd[1]: Started cri-containerd-068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360.scope - libcontainer container 068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360. May 15 15:14:10.434065 containerd[1549]: time="2025-05-15T15:14:10.434008038Z" level=info msg="StartContainer for \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" returns successfully" May 15 15:14:10.642647 containerd[1549]: time="2025-05-15T15:14:10.642494178Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:10.647039 containerd[1549]: time="2025-05-15T15:14:10.646981596Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=77" May 15 15:14:10.649650 containerd[1549]: time="2025-05-15T15:14:10.649592322Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 376.448451ms" May 15 15:14:10.649650 containerd[1549]: time="2025-05-15T15:14:10.649648783Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 15 15:14:10.655949 containerd[1549]: time="2025-05-15T15:14:10.655460679Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" May 15 15:14:10.658212 containerd[1549]: time="2025-05-15T15:14:10.658171914Z" level=info msg="CreateContainer within sandbox \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 15 15:14:10.665871 containerd[1549]: time="2025-05-15T15:14:10.665793048Z" level=info msg="Container d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:10.708328 containerd[1549]: time="2025-05-15T15:14:10.707146799Z" level=info msg="CreateContainer within sandbox \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\"" May 15 15:14:10.713743 containerd[1549]: time="2025-05-15T15:14:10.713660089Z" level=info msg="StartContainer for \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\"" May 15 15:14:10.722187 containerd[1549]: time="2025-05-15T15:14:10.722078098Z" level=info msg="connecting to shim d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118" address="unix:///run/containerd/s/4abf4ae9599c0886af21aef382ac4f41aee7e4460afebcdffc2ce2a9f0037e0a" protocol=ttrpc version=3 May 15 15:14:10.756251 systemd[1]: Started cri-containerd-d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118.scope - libcontainer container d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118. May 15 15:14:10.957120 containerd[1549]: time="2025-05-15T15:14:10.957063730Z" level=info msg="StartContainer for \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" returns successfully" May 15 15:14:10.977331 kubelet[2694]: I0515 15:14:10.977226 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6999544d4f-js9sf" podStartSLOduration=27.229709508 podStartE2EDuration="32.977186991s" podCreationTimestamp="2025-05-15 15:13:38 +0000 UTC" firstStartedPulling="2025-05-15 15:14:04.019353439 +0000 UTC m=+37.700226444" lastFinishedPulling="2025-05-15 15:14:09.766830917 +0000 UTC m=+43.447703927" observedRunningTime="2025-05-15 15:14:10.943615029 +0000 UTC m=+44.624488055" watchObservedRunningTime="2025-05-15 15:14:10.977186991 +0000 UTC m=+44.658060037" May 15 15:14:11.960442 kubelet[2694]: I0515 15:14:11.960402 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:11.979220 kubelet[2694]: I0515 15:14:11.979144 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-585b9b4df4-cr2q4" podStartSLOduration=29.783013572 podStartE2EDuration="33.979115748s" podCreationTimestamp="2025-05-15 15:13:38 +0000 UTC" firstStartedPulling="2025-05-15 15:14:06.076766865 +0000 UTC m=+39.757639872" lastFinishedPulling="2025-05-15 15:14:10.27286904 +0000 UTC m=+43.953742048" observedRunningTime="2025-05-15 15:14:10.981134141 +0000 UTC m=+44.662007194" watchObservedRunningTime="2025-05-15 15:14:11.979115748 +0000 UTC m=+45.659988778" May 15 15:14:12.067296 kubelet[2694]: I0515 15:14:12.066604 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-585b9b4df4-d2bqf" podStartSLOduration=30.51378308 podStartE2EDuration="34.066575966s" podCreationTimestamp="2025-05-15 15:13:38 +0000 UTC" firstStartedPulling="2025-05-15 15:14:07.102513616 +0000 UTC m=+40.783386644" lastFinishedPulling="2025-05-15 15:14:10.655306523 +0000 UTC m=+44.336179530" observedRunningTime="2025-05-15 15:14:11.980524602 +0000 UTC m=+45.661397629" watchObservedRunningTime="2025-05-15 15:14:12.066575966 +0000 UTC m=+45.747448994" May 15 15:14:12.960831 kubelet[2694]: I0515 15:14:12.960747 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:13.617813 containerd[1549]: time="2025-05-15T15:14:13.617021856Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:13.618639 containerd[1549]: time="2025-05-15T15:14:13.618606415Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.3: active requests=0, bytes read=7912898" May 15 15:14:13.619401 containerd[1549]: time="2025-05-15T15:14:13.619363103Z" level=info msg="ImageCreate event name:\"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:13.622346 containerd[1549]: time="2025-05-15T15:14:13.622274396Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:13.623305 containerd[1549]: time="2025-05-15T15:14:13.623267781Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.3\" with image id \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\", size \"9405520\" in 2.967765167s" May 15 15:14:13.623305 containerd[1549]: time="2025-05-15T15:14:13.623304234Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\"" May 15 15:14:13.627994 containerd[1549]: time="2025-05-15T15:14:13.627956791Z" level=info msg="CreateContainer within sandbox \"72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" May 15 15:14:13.645093 containerd[1549]: time="2025-05-15T15:14:13.645045834Z" level=info msg="Container 191164f8a9b9b22e2b43b36b3f35eeaf0bfa3e55e2bcc67e9624e1a23e951aad: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:13.667549 containerd[1549]: time="2025-05-15T15:14:13.667462710Z" level=info msg="CreateContainer within sandbox \"72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"191164f8a9b9b22e2b43b36b3f35eeaf0bfa3e55e2bcc67e9624e1a23e951aad\"" May 15 15:14:13.669559 containerd[1549]: time="2025-05-15T15:14:13.669455502Z" level=info msg="StartContainer for \"191164f8a9b9b22e2b43b36b3f35eeaf0bfa3e55e2bcc67e9624e1a23e951aad\"" May 15 15:14:13.671960 containerd[1549]: time="2025-05-15T15:14:13.671860367Z" level=info msg="connecting to shim 191164f8a9b9b22e2b43b36b3f35eeaf0bfa3e55e2bcc67e9624e1a23e951aad" address="unix:///run/containerd/s/121dcd7c2b44212bf606ebca34fa8945562113b9d906455d304b594015d3efe6" protocol=ttrpc version=3 May 15 15:14:13.718282 systemd[1]: Started cri-containerd-191164f8a9b9b22e2b43b36b3f35eeaf0bfa3e55e2bcc67e9624e1a23e951aad.scope - libcontainer container 191164f8a9b9b22e2b43b36b3f35eeaf0bfa3e55e2bcc67e9624e1a23e951aad. May 15 15:14:13.768214 containerd[1549]: time="2025-05-15T15:14:13.768143109Z" level=info msg="StartContainer for \"191164f8a9b9b22e2b43b36b3f35eeaf0bfa3e55e2bcc67e9624e1a23e951aad\" returns successfully" May 15 15:14:13.771707 containerd[1549]: time="2025-05-15T15:14:13.771653664Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" May 15 15:14:13.990819 kubelet[2694]: I0515 15:14:13.990774 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:14.087451 containerd[1549]: time="2025-05-15T15:14:14.087382486Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" id:\"5bf4011e9d82fc62dd533e331170524deea12d8a81ca842523c571c5ea936458\" pid:4911 exited_at:{seconds:1747322054 nanos:73534382}" May 15 15:14:14.164790 containerd[1549]: time="2025-05-15T15:14:14.164653826Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" id:\"7fb6ad1b1b77ed41a6e9bdd239da3bf4120be4c291edb12638e27f0285360fed\" pid:4933 exited_at:{seconds:1747322054 nanos:163744753}" May 15 15:14:15.155257 systemd[1]: Started sshd@9-24.144.90.91:22-139.178.68.195:53044.service - OpenSSH per-connection server daemon (139.178.68.195:53044). May 15 15:14:15.313917 sshd[4948]: Accepted publickey for core from 139.178.68.195 port 53044 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:15.316787 sshd-session[4948]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:15.328267 systemd-logind[1527]: New session 8 of user core. May 15 15:14:15.334176 systemd[1]: Started session-8.scope - Session 8 of User core. May 15 15:14:15.889269 kubelet[2694]: I0515 15:14:15.888971 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:16.022626 containerd[1549]: time="2025-05-15T15:14:16.022511612Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:16.026057 containerd[1549]: time="2025-05-15T15:14:16.026001390Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3: active requests=0, bytes read=13991773" May 15 15:14:16.028256 containerd[1549]: time="2025-05-15T15:14:16.028148754Z" level=info msg="ImageCreate event name:\"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:16.034961 containerd[1549]: time="2025-05-15T15:14:16.033152044Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 15:14:16.034961 containerd[1549]: time="2025-05-15T15:14:16.034038043Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" with image id \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\", size \"15484347\" in 2.262321833s" May 15 15:14:16.034961 containerd[1549]: time="2025-05-15T15:14:16.034082946Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\"" May 15 15:14:16.042117 containerd[1549]: time="2025-05-15T15:14:16.041808216Z" level=info msg="CreateContainer within sandbox \"72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" May 15 15:14:16.064426 containerd[1549]: time="2025-05-15T15:14:16.064386210Z" level=info msg="Container fca7b9ac2462fe7501beca1acc30cdbf0d30b324c2e911673091321ddbe91dec: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:16.072468 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4175435908.mount: Deactivated successfully. May 15 15:14:16.092301 containerd[1549]: time="2025-05-15T15:14:16.092090836Z" level=info msg="CreateContainer within sandbox \"72b00bdb7986b966d6d865ead5a8b07f35f70beb1e29a9a6f3b421de81bbe409\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"fca7b9ac2462fe7501beca1acc30cdbf0d30b324c2e911673091321ddbe91dec\"" May 15 15:14:16.101959 containerd[1549]: time="2025-05-15T15:14:16.100120029Z" level=info msg="StartContainer for \"fca7b9ac2462fe7501beca1acc30cdbf0d30b324c2e911673091321ddbe91dec\"" May 15 15:14:16.103786 sshd[4950]: Connection closed by 139.178.68.195 port 53044 May 15 15:14:16.103554 sshd-session[4948]: pam_unix(sshd:session): session closed for user core May 15 15:14:16.107401 containerd[1549]: time="2025-05-15T15:14:16.106475880Z" level=info msg="connecting to shim fca7b9ac2462fe7501beca1acc30cdbf0d30b324c2e911673091321ddbe91dec" address="unix:///run/containerd/s/121dcd7c2b44212bf606ebca34fa8945562113b9d906455d304b594015d3efe6" protocol=ttrpc version=3 May 15 15:14:16.135109 systemd[1]: sshd@9-24.144.90.91:22-139.178.68.195:53044.service: Deactivated successfully. May 15 15:14:16.141761 systemd[1]: session-8.scope: Deactivated successfully. May 15 15:14:16.153837 systemd-logind[1527]: Session 8 logged out. Waiting for processes to exit. May 15 15:14:16.158641 systemd-logind[1527]: Removed session 8. May 15 15:14:16.181993 systemd[1]: Started cri-containerd-fca7b9ac2462fe7501beca1acc30cdbf0d30b324c2e911673091321ddbe91dec.scope - libcontainer container fca7b9ac2462fe7501beca1acc30cdbf0d30b324c2e911673091321ddbe91dec. May 15 15:14:16.278095 containerd[1549]: time="2025-05-15T15:14:16.278050130Z" level=info msg="StopContainer for \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" with timeout 30 (s)" May 15 15:14:16.281562 systemd[1]: Created slice kubepods-besteffort-poddea30fea_d7ce_49f3_8528_5f0c490f48fa.slice - libcontainer container kubepods-besteffort-poddea30fea_d7ce_49f3_8528_5f0c490f48fa.slice. May 15 15:14:16.289273 containerd[1549]: time="2025-05-15T15:14:16.289233639Z" level=info msg="Stop container \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" with signal terminated" May 15 15:14:16.309150 systemd[1]: cri-containerd-d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118.scope: Deactivated successfully. May 15 15:14:16.315570 containerd[1549]: time="2025-05-15T15:14:16.315477306Z" level=info msg="received exit event container_id:\"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" id:\"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" pid:4832 exit_status:1 exited_at:{seconds:1747322056 nanos:315124641}" May 15 15:14:16.317231 containerd[1549]: time="2025-05-15T15:14:16.316391451Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" id:\"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" pid:4832 exit_status:1 exited_at:{seconds:1747322056 nanos:315124641}" May 15 15:14:16.349440 kubelet[2694]: I0515 15:14:16.349276 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/dea30fea-d7ce-49f3-8528-5f0c490f48fa-calico-apiserver-certs\") pod \"calico-apiserver-6999544d4f-77ccn\" (UID: \"dea30fea-d7ce-49f3-8528-5f0c490f48fa\") " pod="calico-apiserver/calico-apiserver-6999544d4f-77ccn" May 15 15:14:16.349851 kubelet[2694]: I0515 15:14:16.349604 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmvgv\" (UniqueName: \"kubernetes.io/projected/dea30fea-d7ce-49f3-8528-5f0c490f48fa-kube-api-access-nmvgv\") pod \"calico-apiserver-6999544d4f-77ccn\" (UID: \"dea30fea-d7ce-49f3-8528-5f0c490f48fa\") " pod="calico-apiserver/calico-apiserver-6999544d4f-77ccn" May 15 15:14:16.350393 containerd[1549]: time="2025-05-15T15:14:16.350353885Z" level=info msg="StartContainer for \"fca7b9ac2462fe7501beca1acc30cdbf0d30b324c2e911673091321ddbe91dec\" returns successfully" May 15 15:14:16.383656 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118-rootfs.mount: Deactivated successfully. May 15 15:14:16.410112 containerd[1549]: time="2025-05-15T15:14:16.409564092Z" level=info msg="StopContainer for \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" returns successfully" May 15 15:14:16.429034 containerd[1549]: time="2025-05-15T15:14:16.428996250Z" level=info msg="StopPodSandbox for \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\"" May 15 15:14:16.429328 containerd[1549]: time="2025-05-15T15:14:16.429304742Z" level=info msg="Container to stop \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 15 15:14:16.438192 systemd[1]: cri-containerd-5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30.scope: Deactivated successfully. May 15 15:14:16.440687 containerd[1549]: time="2025-05-15T15:14:16.440535690Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" id:\"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" pid:4653 exit_status:137 exited_at:{seconds:1747322056 nanos:440206948}" May 15 15:14:16.512717 containerd[1549]: time="2025-05-15T15:14:16.512610565Z" level=info msg="shim disconnected" id=5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30 namespace=k8s.io May 15 15:14:16.512717 containerd[1549]: time="2025-05-15T15:14:16.512647420Z" level=warning msg="cleaning up after shim disconnected" id=5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30 namespace=k8s.io May 15 15:14:16.523814 containerd[1549]: time="2025-05-15T15:14:16.512657165Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 15 15:14:16.596974 containerd[1549]: time="2025-05-15T15:14:16.596902853Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6999544d4f-77ccn,Uid:dea30fea-d7ce-49f3-8528-5f0c490f48fa,Namespace:calico-apiserver,Attempt:0,}" May 15 15:14:16.601965 containerd[1549]: time="2025-05-15T15:14:16.601215573Z" level=info msg="received exit event sandbox_id:\"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" exit_status:137 exited_at:{seconds:1747322056 nanos:440206948}" May 15 15:14:16.841454 systemd-networkd[1455]: caliabce917082c: Link DOWN May 15 15:14:16.841464 systemd-networkd[1455]: caliabce917082c: Lost carrier May 15 15:14:16.877686 kubelet[2694]: I0515 15:14:16.877543 2694 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 May 15 15:14:16.878109 kubelet[2694]: I0515 15:14:16.877707 2694 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock May 15 15:14:17.036694 systemd-networkd[1455]: calieaf53c78784: Link UP May 15 15:14:17.037045 systemd-networkd[1455]: calieaf53c78784: Gained carrier May 15 15:14:17.079411 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30-rootfs.mount: Deactivated successfully. May 15 15:14:17.079592 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30-shm.mount: Deactivated successfully. May 15 15:14:17.109911 kubelet[2694]: I0515 15:14:17.109390 2694 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.744 [INFO][5084] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0 calico-apiserver-6999544d4f- calico-apiserver dea30fea-d7ce-49f3-8528-5f0c490f48fa 1013 0 2025-05-15 15:14:16 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6999544d4f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a calico-apiserver-6999544d4f-77ccn eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calieaf53c78784 [] []}} ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-77ccn" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.746 [INFO][5084] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-77ccn" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.880 [INFO][5104] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" HandleID="k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.905 [INFO][5104] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" HandleID="k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00039fb90), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4334.0.0-a-073c24674a", "pod":"calico-apiserver-6999544d4f-77ccn", "timestamp":"2025-05-15 15:14:16.880143328 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.908 [INFO][5104] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.909 [INFO][5104] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.909 [INFO][5104] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.920 [INFO][5104] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.931 [INFO][5104] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.943 [INFO][5104] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.952 [INFO][5104] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.959 [INFO][5104] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.960 [INFO][5104] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.964 [INFO][5104] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7 May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.971 [INFO][5104] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.994 [INFO][5104] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.72/26] block=192.168.50.64/26 handle="k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.994 [INFO][5104] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.72/26] handle="k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" host="ci-4334.0.0-a-073c24674a" May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.994 [INFO][5104] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:17.112695 containerd[1549]: 2025-05-15 15:14:16.994 [INFO][5104] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.72/26] IPv6=[] ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" HandleID="k8s-pod-network.170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" May 15 15:14:17.114318 containerd[1549]: 2025-05-15 15:14:17.010 [INFO][5084] cni-plugin/k8s.go 386: Populated endpoint ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-77ccn" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0", GenerateName:"calico-apiserver-6999544d4f-", Namespace:"calico-apiserver", SelfLink:"", UID:"dea30fea-d7ce-49f3-8528-5f0c490f48fa", ResourceVersion:"1013", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 14, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6999544d4f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"calico-apiserver-6999544d4f-77ccn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.50.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieaf53c78784", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:17.114318 containerd[1549]: 2025-05-15 15:14:17.011 [INFO][5084] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.72/32] ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-77ccn" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" May 15 15:14:17.114318 containerd[1549]: 2025-05-15 15:14:17.011 [INFO][5084] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieaf53c78784 ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-77ccn" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" May 15 15:14:17.114318 containerd[1549]: 2025-05-15 15:14:17.037 [INFO][5084] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-77ccn" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" May 15 15:14:17.114318 containerd[1549]: 2025-05-15 15:14:17.037 [INFO][5084] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-77ccn" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0", GenerateName:"calico-apiserver-6999544d4f-", Namespace:"calico-apiserver", SelfLink:"", UID:"dea30fea-d7ce-49f3-8528-5f0c490f48fa", ResourceVersion:"1013", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 14, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6999544d4f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7", Pod:"calico-apiserver-6999544d4f-77ccn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.50.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieaf53c78784", MAC:"da:f3:5a:71:d8:a8", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:17.114318 containerd[1549]: 2025-05-15 15:14:17.068 [INFO][5084] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" Namespace="calico-apiserver" Pod="calico-apiserver-6999544d4f-77ccn" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--6999544d4f--77ccn-eth0" May 15 15:14:17.143685 kubelet[2694]: I0515 15:14:17.143426 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-5mw49" podStartSLOduration=30.936129241 podStartE2EDuration="39.14340344s" podCreationTimestamp="2025-05-15 15:13:38 +0000 UTC" firstStartedPulling="2025-05-15 15:14:07.830839917 +0000 UTC m=+41.511712923" lastFinishedPulling="2025-05-15 15:14:16.038114116 +0000 UTC m=+49.718987122" observedRunningTime="2025-05-15 15:14:17.138901794 +0000 UTC m=+50.819774837" watchObservedRunningTime="2025-05-15 15:14:17.14340344 +0000 UTC m=+50.824276466" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.832 [INFO][5091] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.834 [INFO][5091] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" iface="eth0" netns="/var/run/netns/cni-b199727c-0c89-2699-18c4-741eeab967d7" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.835 [INFO][5091] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" iface="eth0" netns="/var/run/netns/cni-b199727c-0c89-2699-18c4-741eeab967d7" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.851 [INFO][5091] cni-plugin/dataplane_linux.go 604: Deleted device in netns. ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" after=16.717844ms iface="eth0" netns="/var/run/netns/cni-b199727c-0c89-2699-18c4-741eeab967d7" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.856 [INFO][5091] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.856 [INFO][5091] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.941 [INFO][5112] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.943 [INFO][5112] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:16.994 [INFO][5112] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:17.161 [INFO][5112] ipam/ipam_plugin.go 431: Released address using handleID ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:17.165 [INFO][5112] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:17.173 [INFO][5112] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:17.194901 containerd[1549]: 2025-05-15 15:14:17.183 [INFO][5091] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:17.204478 containerd[1549]: time="2025-05-15T15:14:17.204288242Z" level=info msg="connecting to shim 170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7" address="unix:///run/containerd/s/c9a87181422d2f0a2eb177d7eb22602b5cb9b0eef0aca7b268d06e31b61f707a" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:17.209076 containerd[1549]: time="2025-05-15T15:14:17.209013994Z" level=info msg="TearDown network for sandbox \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" successfully" May 15 15:14:17.209076 containerd[1549]: time="2025-05-15T15:14:17.209071866Z" level=info msg="StopPodSandbox for \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" returns successfully" May 15 15:14:17.211618 systemd[1]: run-netns-cni\x2db199727c\x2d0c89\x2d2699\x2d18c4\x2d741eeab967d7.mount: Deactivated successfully. May 15 15:14:17.296229 systemd[1]: Started cri-containerd-170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7.scope - libcontainer container 170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7. May 15 15:14:17.373868 kubelet[2694]: I0515 15:14:17.373429 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2brf\" (UniqueName: \"kubernetes.io/projected/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7-kube-api-access-q2brf\") pod \"1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7\" (UID: \"1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7\") " May 15 15:14:17.373868 kubelet[2694]: I0515 15:14:17.373498 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7-calico-apiserver-certs\") pod \"1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7\" (UID: \"1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7\") " May 15 15:14:17.395923 systemd[1]: var-lib-kubelet-pods-1c80ddf0\x2de33c\x2d4576\x2d8bd9\x2d03a8cc5f33e7-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dq2brf.mount: Deactivated successfully. May 15 15:14:17.397312 kubelet[2694]: I0515 15:14:17.395616 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7-kube-api-access-q2brf" (OuterVolumeSpecName: "kube-api-access-q2brf") pod "1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7" (UID: "1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7"). InnerVolumeSpecName "kube-api-access-q2brf". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 15 15:14:17.401162 kubelet[2694]: I0515 15:14:17.401112 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7-calico-apiserver-certs" (OuterVolumeSpecName: "calico-apiserver-certs") pod "1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7" (UID: "1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7"). InnerVolumeSpecName "calico-apiserver-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 15 15:14:17.444858 containerd[1549]: time="2025-05-15T15:14:17.444773050Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6999544d4f-77ccn,Uid:dea30fea-d7ce-49f3-8528-5f0c490f48fa,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7\"" May 15 15:14:17.458958 containerd[1549]: time="2025-05-15T15:14:17.458892939Z" level=info msg="CreateContainer within sandbox \"170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 15 15:14:17.472378 containerd[1549]: time="2025-05-15T15:14:17.472286921Z" level=info msg="Container ef7e7182a86203c2781b94e032a42c67b05b53fbe5bdf25d9a247136770156a0: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:17.474493 kubelet[2694]: I0515 15:14:17.474389 2694 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-q2brf\" (UniqueName: \"kubernetes.io/projected/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7-kube-api-access-q2brf\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:17.474493 kubelet[2694]: I0515 15:14:17.474462 2694 reconciler_common.go:299] "Volume detached for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7-calico-apiserver-certs\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:17.484469 containerd[1549]: time="2025-05-15T15:14:17.484424865Z" level=info msg="CreateContainer within sandbox \"170322ba9a376e60b127e94cb445cdd309b329ee6a10b4eca6114439da5537d7\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"ef7e7182a86203c2781b94e032a42c67b05b53fbe5bdf25d9a247136770156a0\"" May 15 15:14:17.486969 containerd[1549]: time="2025-05-15T15:14:17.486939875Z" level=info msg="StartContainer for \"ef7e7182a86203c2781b94e032a42c67b05b53fbe5bdf25d9a247136770156a0\"" May 15 15:14:17.488434 containerd[1549]: time="2025-05-15T15:14:17.488354963Z" level=info msg="connecting to shim ef7e7182a86203c2781b94e032a42c67b05b53fbe5bdf25d9a247136770156a0" address="unix:///run/containerd/s/c9a87181422d2f0a2eb177d7eb22602b5cb9b0eef0aca7b268d06e31b61f707a" protocol=ttrpc version=3 May 15 15:14:17.518330 systemd[1]: Started cri-containerd-ef7e7182a86203c2781b94e032a42c67b05b53fbe5bdf25d9a247136770156a0.scope - libcontainer container ef7e7182a86203c2781b94e032a42c67b05b53fbe5bdf25d9a247136770156a0. May 15 15:14:17.629084 containerd[1549]: time="2025-05-15T15:14:17.626680736Z" level=info msg="StartContainer for \"ef7e7182a86203c2781b94e032a42c67b05b53fbe5bdf25d9a247136770156a0\" returns successfully" May 15 15:14:18.075149 systemd[1]: var-lib-kubelet-pods-1c80ddf0\x2de33c\x2d4576\x2d8bd9\x2d03a8cc5f33e7-volumes-kubernetes.io\x7esecret-calico\x2dapiserver\x2dcerts.mount: Deactivated successfully. May 15 15:14:18.087797 systemd[1]: Removed slice kubepods-besteffort-pod1c80ddf0_e33c_4576_8bd9_03a8cc5f33e7.slice - libcontainer container kubepods-besteffort-pod1c80ddf0_e33c_4576_8bd9_03a8cc5f33e7.slice. May 15 15:14:18.120809 kubelet[2694]: I0515 15:14:18.120752 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6999544d4f-77ccn" podStartSLOduration=2.120730888 podStartE2EDuration="2.120730888s" podCreationTimestamp="2025-05-15 15:14:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:14:18.100218693 +0000 UTC m=+51.781091719" watchObservedRunningTime="2025-05-15 15:14:18.120730888 +0000 UTC m=+51.801603959" May 15 15:14:18.489483 kubelet[2694]: I0515 15:14:18.489167 2694 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7" path="/var/lib/kubelet/pods/1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7/volumes" May 15 15:14:18.767171 systemd-networkd[1455]: calieaf53c78784: Gained IPv6LL May 15 15:14:19.074739 kubelet[2694]: I0515 15:14:19.074198 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:20.207922 containerd[1549]: time="2025-05-15T15:14:20.207874544Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" id:\"1dec45418a2739549720e51ffa7964f2b8ed9490aa0b363eefadbecd9b86fcae\" pid:5245 exited_at:{seconds:1747322060 nanos:207227935}" May 15 15:14:21.126028 systemd[1]: Started sshd@10-24.144.90.91:22-139.178.68.195:53050.service - OpenSSH per-connection server daemon (139.178.68.195:53050). May 15 15:14:21.213474 sshd[5256]: Accepted publickey for core from 139.178.68.195 port 53050 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:21.215960 sshd-session[5256]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:21.225082 systemd-logind[1527]: New session 9 of user core. May 15 15:14:21.230222 systemd[1]: Started session-9.scope - Session 9 of User core. May 15 15:14:21.431990 sshd[5261]: Connection closed by 139.178.68.195 port 53050 May 15 15:14:21.432689 sshd-session[5256]: pam_unix(sshd:session): session closed for user core May 15 15:14:21.436708 systemd-logind[1527]: Session 9 logged out. Waiting for processes to exit. May 15 15:14:21.436870 systemd[1]: sshd@10-24.144.90.91:22-139.178.68.195:53050.service: Deactivated successfully. May 15 15:14:21.439168 systemd[1]: session-9.scope: Deactivated successfully. May 15 15:14:21.442730 systemd-logind[1527]: Removed session 9. May 15 15:14:26.449183 systemd[1]: Started sshd@11-24.144.90.91:22-139.178.68.195:43322.service - OpenSSH per-connection server daemon (139.178.68.195:43322). May 15 15:14:26.503089 kubelet[2694]: I0515 15:14:26.503047 2694 scope.go:117] "RemoveContainer" containerID="d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118" May 15 15:14:26.508200 containerd[1549]: time="2025-05-15T15:14:26.507756725Z" level=info msg="RemoveContainer for \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\"" May 15 15:14:26.525241 containerd[1549]: time="2025-05-15T15:14:26.525079723Z" level=info msg="RemoveContainer for \"d2a186a1e8ff90c2bdd5a04a7e6a0b29c263eae9a802ec337e7ca039010fb118\" returns successfully" May 15 15:14:26.530002 containerd[1549]: time="2025-05-15T15:14:26.529960995Z" level=info msg="StopPodSandbox for \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\"" May 15 15:14:26.559911 sshd[5282]: Accepted publickey for core from 139.178.68.195 port 43322 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:26.561197 sshd-session[5282]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:26.569716 systemd-logind[1527]: New session 10 of user core. May 15 15:14:26.577674 systemd[1]: Started session-10.scope - Session 10 of User core. May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.600 [WARNING][5300] cni-plugin/k8s.go 566: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.600 [INFO][5300] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.600 [INFO][5300] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" iface="eth0" netns="" May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.600 [INFO][5300] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.601 [INFO][5300] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.652 [INFO][5308] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.652 [INFO][5308] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.652 [INFO][5308] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.665 [WARNING][5308] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.665 [INFO][5308] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.669 [INFO][5308] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:26.677368 containerd[1549]: 2025-05-15 15:14:26.673 [INFO][5300] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:26.678262 containerd[1549]: time="2025-05-15T15:14:26.677419571Z" level=info msg="TearDown network for sandbox \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" successfully" May 15 15:14:26.678262 containerd[1549]: time="2025-05-15T15:14:26.677446149Z" level=info msg="StopPodSandbox for \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" returns successfully" May 15 15:14:26.679752 containerd[1549]: time="2025-05-15T15:14:26.679649363Z" level=info msg="RemovePodSandbox for \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\"" May 15 15:14:26.679752 containerd[1549]: time="2025-05-15T15:14:26.679719119Z" level=info msg="Forcibly stopping sandbox \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\"" May 15 15:14:26.834612 kubelet[2694]: I0515 15:14:26.831630 2694 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 15:14:26.848885 sshd[5305]: Connection closed by 139.178.68.195 port 43322 May 15 15:14:26.858429 sshd-session[5282]: pam_unix(sshd:session): session closed for user core May 15 15:14:26.872863 systemd[1]: sshd@11-24.144.90.91:22-139.178.68.195:43322.service: Deactivated successfully. May 15 15:14:26.875961 systemd[1]: session-10.scope: Deactivated successfully. May 15 15:14:26.877247 systemd-logind[1527]: Session 10 logged out. Waiting for processes to exit. May 15 15:14:26.884176 systemd[1]: Started sshd@12-24.144.90.91:22-139.178.68.195:43330.service - OpenSSH per-connection server daemon (139.178.68.195:43330). May 15 15:14:26.885499 systemd-logind[1527]: Removed session 10. May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.802 [WARNING][5334] cni-plugin/k8s.go 566: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.803 [INFO][5334] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.804 [INFO][5334] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" iface="eth0" netns="" May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.804 [INFO][5334] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.804 [INFO][5334] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.865 [INFO][5341] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.865 [INFO][5341] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.868 [INFO][5341] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.903 [WARNING][5341] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.903 [INFO][5341] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" HandleID="k8s-pod-network.5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--d2bqf-eth0" May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.916 [INFO][5341] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:26.934704 containerd[1549]: 2025-05-15 15:14:26.924 [INFO][5334] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30" May 15 15:14:26.934704 containerd[1549]: time="2025-05-15T15:14:26.934580728Z" level=info msg="TearDown network for sandbox \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" successfully" May 15 15:14:26.941322 containerd[1549]: time="2025-05-15T15:14:26.941272284Z" level=info msg="Ensure that sandbox 5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30 in task-service has been cleanup successfully" May 15 15:14:26.945954 containerd[1549]: time="2025-05-15T15:14:26.945401674Z" level=info msg="RemovePodSandbox \"5e52a2564c0dce4fdab4b35ea1d68be09065aca58a142488973be8a5fb59ab30\" returns successfully" May 15 15:14:26.977769 containerd[1549]: time="2025-05-15T15:14:26.977729497Z" level=info msg="StopContainer for \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" with timeout 30 (s)" May 15 15:14:26.983332 containerd[1549]: time="2025-05-15T15:14:26.983278137Z" level=info msg="Stop container \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" with signal terminated" May 15 15:14:27.019598 sshd[5352]: Accepted publickey for core from 139.178.68.195 port 43330 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:27.021417 sshd-session[5352]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:27.021493 systemd[1]: cri-containerd-068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360.scope: Deactivated successfully. May 15 15:14:27.030568 containerd[1549]: time="2025-05-15T15:14:27.030362293Z" level=info msg="received exit event container_id:\"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" id:\"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" pid:4796 exit_status:1 exited_at:{seconds:1747322067 nanos:27852049}" May 15 15:14:27.032197 containerd[1549]: time="2025-05-15T15:14:27.031991600Z" level=info msg="TaskExit event in podsandbox handler container_id:\"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" id:\"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" pid:4796 exit_status:1 exited_at:{seconds:1747322067 nanos:27852049}" May 15 15:14:27.032781 systemd-logind[1527]: New session 11 of user core. May 15 15:14:27.037378 systemd[1]: Started session-11.scope - Session 11 of User core. May 15 15:14:27.070112 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360-rootfs.mount: Deactivated successfully. May 15 15:14:27.079171 containerd[1549]: time="2025-05-15T15:14:27.079043833Z" level=info msg="StopContainer for \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" returns successfully" May 15 15:14:27.080114 containerd[1549]: time="2025-05-15T15:14:27.080075849Z" level=info msg="StopPodSandbox for \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\"" May 15 15:14:27.080239 containerd[1549]: time="2025-05-15T15:14:27.080184729Z" level=info msg="Container to stop \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 15 15:14:27.093295 systemd[1]: cri-containerd-41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611.scope: Deactivated successfully. May 15 15:14:27.102089 containerd[1549]: time="2025-05-15T15:14:27.102044401Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" id:\"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" pid:4543 exit_status:137 exited_at:{seconds:1747322067 nanos:99363351}" May 15 15:14:27.184580 containerd[1549]: time="2025-05-15T15:14:27.184533257Z" level=info msg="shim disconnected" id=41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611 namespace=k8s.io May 15 15:14:27.184580 containerd[1549]: time="2025-05-15T15:14:27.184565366Z" level=warning msg="cleaning up after shim disconnected" id=41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611 namespace=k8s.io May 15 15:14:27.184746 containerd[1549]: time="2025-05-15T15:14:27.184575779Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 15 15:14:27.185458 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611-rootfs.mount: Deactivated successfully. May 15 15:14:27.236579 containerd[1549]: time="2025-05-15T15:14:27.236300745Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" id:\"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" pid:4543 exit_status:137 exited_at:{seconds:1747322067 nanos:225386311}" May 15 15:14:27.236579 containerd[1549]: time="2025-05-15T15:14:27.236435882Z" level=info msg="received exit event sandbox_id:\"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" exit_status:137 exited_at:{seconds:1747322067 nanos:225386311}" May 15 15:14:27.240755 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611-shm.mount: Deactivated successfully. May 15 15:14:27.387195 systemd-networkd[1455]: cali743a39b7768: Link DOWN May 15 15:14:27.387203 systemd-networkd[1455]: cali743a39b7768: Lost carrier May 15 15:14:27.393032 sshd[5366]: Connection closed by 139.178.68.195 port 43330 May 15 15:14:27.392369 sshd-session[5352]: pam_unix(sshd:session): session closed for user core May 15 15:14:27.415502 systemd[1]: sshd@12-24.144.90.91:22-139.178.68.195:43330.service: Deactivated successfully. May 15 15:14:27.422488 systemd[1]: session-11.scope: Deactivated successfully. May 15 15:14:27.425884 systemd-logind[1527]: Session 11 logged out. Waiting for processes to exit. May 15 15:14:27.435345 systemd[1]: Started sshd@13-24.144.90.91:22-139.178.68.195:43340.service - OpenSSH per-connection server daemon (139.178.68.195:43340). May 15 15:14:27.438764 systemd-logind[1527]: Removed session 11. May 15 15:14:27.517204 sshd[5455]: Accepted publickey for core from 139.178.68.195 port 43340 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:27.519423 sshd-session[5455]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:27.531094 systemd-logind[1527]: New session 12 of user core. May 15 15:14:27.536238 systemd[1]: Started session-12.scope - Session 12 of User core. May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.385 [INFO][5438] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.385 [INFO][5438] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" iface="eth0" netns="/var/run/netns/cni-31de5996-a100-6ada-8921-25d199e9ce3c" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.385 [INFO][5438] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" iface="eth0" netns="/var/run/netns/cni-31de5996-a100-6ada-8921-25d199e9ce3c" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.400 [INFO][5438] cni-plugin/dataplane_linux.go 604: Deleted device in netns. ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" after=14.950302ms iface="eth0" netns="/var/run/netns/cni-31de5996-a100-6ada-8921-25d199e9ce3c" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.400 [INFO][5438] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.400 [INFO][5438] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.507 [INFO][5446] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.507 [INFO][5446] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.507 [INFO][5446] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.572 [INFO][5446] ipam/ipam_plugin.go 431: Released address using handleID ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.572 [INFO][5446] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.574 [INFO][5446] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:27.582418 containerd[1549]: 2025-05-15 15:14:27.578 [INFO][5438] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:14:27.586109 containerd[1549]: time="2025-05-15T15:14:27.586062662Z" level=info msg="TearDown network for sandbox \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" successfully" May 15 15:14:27.586109 containerd[1549]: time="2025-05-15T15:14:27.586097181Z" level=info msg="StopPodSandbox for \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" returns successfully" May 15 15:14:27.587122 systemd[1]: run-netns-cni\x2d31de5996\x2da100\x2d6ada\x2d8921\x2d25d199e9ce3c.mount: Deactivated successfully. May 15 15:14:27.716592 sshd[5463]: Connection closed by 139.178.68.195 port 43340 May 15 15:14:27.717061 sshd-session[5455]: pam_unix(sshd:session): session closed for user core May 15 15:14:27.725908 systemd-logind[1527]: Session 12 logged out. Waiting for processes to exit. May 15 15:14:27.726558 systemd[1]: sshd@13-24.144.90.91:22-139.178.68.195:43340.service: Deactivated successfully. May 15 15:14:27.730317 systemd[1]: session-12.scope: Deactivated successfully. May 15 15:14:27.734651 systemd-logind[1527]: Removed session 12. May 15 15:14:27.756855 kubelet[2694]: I0515 15:14:27.756816 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktx9g\" (UniqueName: \"kubernetes.io/projected/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad-kube-api-access-ktx9g\") pod \"52c8dc92-1a03-4e1d-bc5b-3844d9a477ad\" (UID: \"52c8dc92-1a03-4e1d-bc5b-3844d9a477ad\") " May 15 15:14:27.756855 kubelet[2694]: I0515 15:14:27.756864 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad-calico-apiserver-certs\") pod \"52c8dc92-1a03-4e1d-bc5b-3844d9a477ad\" (UID: \"52c8dc92-1a03-4e1d-bc5b-3844d9a477ad\") " May 15 15:14:27.765804 kubelet[2694]: I0515 15:14:27.765720 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad-kube-api-access-ktx9g" (OuterVolumeSpecName: "kube-api-access-ktx9g") pod "52c8dc92-1a03-4e1d-bc5b-3844d9a477ad" (UID: "52c8dc92-1a03-4e1d-bc5b-3844d9a477ad"). InnerVolumeSpecName "kube-api-access-ktx9g". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 15 15:14:27.770050 systemd[1]: var-lib-kubelet-pods-52c8dc92\x2d1a03\x2d4e1d\x2dbc5b\x2d3844d9a477ad-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dktx9g.mount: Deactivated successfully. May 15 15:14:27.771410 kubelet[2694]: I0515 15:14:27.771323 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad-calico-apiserver-certs" (OuterVolumeSpecName: "calico-apiserver-certs") pod "52c8dc92-1a03-4e1d-bc5b-3844d9a477ad" (UID: "52c8dc92-1a03-4e1d-bc5b-3844d9a477ad"). InnerVolumeSpecName "calico-apiserver-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 15 15:14:27.857239 kubelet[2694]: I0515 15:14:27.857198 2694 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ktx9g\" (UniqueName: \"kubernetes.io/projected/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad-kube-api-access-ktx9g\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:27.857239 kubelet[2694]: I0515 15:14:27.857235 2694 reconciler_common.go:299] "Volume detached for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad-calico-apiserver-certs\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:28.070448 systemd[1]: var-lib-kubelet-pods-52c8dc92\x2d1a03\x2d4e1d\x2dbc5b\x2d3844d9a477ad-volumes-kubernetes.io\x7esecret-calico\x2dapiserver\x2dcerts.mount: Deactivated successfully. May 15 15:14:28.121312 kubelet[2694]: I0515 15:14:28.121216 2694 scope.go:117] "RemoveContainer" containerID="068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360" May 15 15:14:28.124052 containerd[1549]: time="2025-05-15T15:14:28.123964405Z" level=info msg="RemoveContainer for \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\"" May 15 15:14:28.129155 containerd[1549]: time="2025-05-15T15:14:28.128594236Z" level=info msg="RemoveContainer for \"068a9ee8a246b5fa3bc9570a04a7e0ed7a74fb1796becc5255dc71060a618360\" returns successfully" May 15 15:14:28.129523 systemd[1]: Removed slice kubepods-besteffort-pod52c8dc92_1a03_4e1d_bc5b_3844d9a477ad.slice - libcontainer container kubepods-besteffort-pod52c8dc92_1a03_4e1d_bc5b_3844d9a477ad.slice. May 15 15:14:28.485757 kubelet[2694]: I0515 15:14:28.485498 2694 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52c8dc92-1a03-4e1d-bc5b-3844d9a477ad" path="/var/lib/kubelet/pods/52c8dc92-1a03-4e1d-bc5b-3844d9a477ad/volumes" May 15 15:14:28.782398 containerd[1549]: time="2025-05-15T15:14:28.782241449Z" level=info msg="TaskExit event in podsandbox handler exit_status:137 exited_at:{seconds:1747322067 nanos:99363351}" May 15 15:14:31.849179 containerd[1549]: time="2025-05-15T15:14:31.849107543Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" id:\"ab65aafc50c052ba27759e055e2cc999f8d6c6b2938770ddabb13d27f9335613\" pid:5491 exited_at:{seconds:1747322071 nanos:848327097}" May 15 15:14:31.854431 kubelet[2694]: E0515 15:14:31.853182 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:32.734919 systemd[1]: Started sshd@14-24.144.90.91:22-139.178.68.195:43344.service - OpenSSH per-connection server daemon (139.178.68.195:43344). May 15 15:14:32.808901 sshd[5508]: Accepted publickey for core from 139.178.68.195 port 43344 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:32.810740 sshd-session[5508]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:32.820076 systemd-logind[1527]: New session 13 of user core. May 15 15:14:32.825443 systemd[1]: Started session-13.scope - Session 13 of User core. May 15 15:14:33.047445 sshd[5510]: Connection closed by 139.178.68.195 port 43344 May 15 15:14:33.048409 sshd-session[5508]: pam_unix(sshd:session): session closed for user core May 15 15:14:33.057388 systemd[1]: sshd@14-24.144.90.91:22-139.178.68.195:43344.service: Deactivated successfully. May 15 15:14:33.061790 systemd[1]: session-13.scope: Deactivated successfully. May 15 15:14:33.065696 systemd-logind[1527]: Session 13 logged out. Waiting for processes to exit. May 15 15:14:33.070689 systemd-logind[1527]: Removed session 13. May 15 15:14:35.732101 systemd[1]: Started sshd@15-24.144.90.91:22-139.150.69.56:57088.service - OpenSSH per-connection server daemon (139.150.69.56:57088). May 15 15:14:36.636807 sshd[5527]: Received disconnect from 139.150.69.56 port 57088:11: Bye Bye [preauth] May 15 15:14:36.636807 sshd[5527]: Disconnected from authenticating user root 139.150.69.56 port 57088 [preauth] May 15 15:14:36.640332 systemd[1]: sshd@15-24.144.90.91:22-139.150.69.56:57088.service: Deactivated successfully. May 15 15:14:38.063596 systemd[1]: Started sshd@16-24.144.90.91:22-139.178.68.195:48066.service - OpenSSH per-connection server daemon (139.178.68.195:48066). May 15 15:14:38.130289 sshd[5535]: Accepted publickey for core from 139.178.68.195 port 48066 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:38.132275 sshd-session[5535]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:38.139797 systemd-logind[1527]: New session 14 of user core. May 15 15:14:38.144123 systemd[1]: Started session-14.scope - Session 14 of User core. May 15 15:14:38.282009 sshd[5537]: Connection closed by 139.178.68.195 port 48066 May 15 15:14:38.284169 sshd-session[5535]: pam_unix(sshd:session): session closed for user core May 15 15:14:38.289094 systemd[1]: sshd@16-24.144.90.91:22-139.178.68.195:48066.service: Deactivated successfully. May 15 15:14:38.292813 systemd[1]: session-14.scope: Deactivated successfully. May 15 15:14:38.293729 systemd-logind[1527]: Session 14 logged out. Waiting for processes to exit. May 15 15:14:38.295789 systemd-logind[1527]: Removed session 14. May 15 15:14:38.948219 containerd[1549]: time="2025-05-15T15:14:38.948159496Z" level=info msg="StopContainer for \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" with timeout 300 (s)" May 15 15:14:38.949023 containerd[1549]: time="2025-05-15T15:14:38.948905391Z" level=info msg="Stop container \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" with signal terminated" May 15 15:14:39.167306 containerd[1549]: time="2025-05-15T15:14:39.167091017Z" level=info msg="StopContainer for \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" with timeout 30 (s)" May 15 15:14:39.168086 containerd[1549]: time="2025-05-15T15:14:39.167924277Z" level=info msg="Stop container \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" with signal terminated" May 15 15:14:39.199113 systemd[1]: cri-containerd-6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11.scope: Deactivated successfully. May 15 15:14:39.199491 systemd[1]: cri-containerd-6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11.scope: Consumed 309ms CPU time, 16.3M memory peak, 2.2M read from disk, 20K written to disk. May 15 15:14:39.203395 containerd[1549]: time="2025-05-15T15:14:39.202886424Z" level=info msg="received exit event container_id:\"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" id:\"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" pid:4584 exit_status:2 exited_at:{seconds:1747322079 nanos:202610026}" May 15 15:14:39.205226 containerd[1549]: time="2025-05-15T15:14:39.205087170Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" id:\"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" pid:4584 exit_status:2 exited_at:{seconds:1747322079 nanos:202610026}" May 15 15:14:39.260236 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11-rootfs.mount: Deactivated successfully. May 15 15:14:39.268396 containerd[1549]: time="2025-05-15T15:14:39.268350689Z" level=info msg="StopContainer for \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" returns successfully" May 15 15:14:39.269374 containerd[1549]: time="2025-05-15T15:14:39.269336808Z" level=info msg="StopPodSandbox for \"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\"" May 15 15:14:39.269592 containerd[1549]: time="2025-05-15T15:14:39.269428747Z" level=info msg="Container to stop \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 15 15:14:39.287441 systemd[1]: cri-containerd-a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f.scope: Deactivated successfully. May 15 15:14:39.291872 containerd[1549]: time="2025-05-15T15:14:39.291776309Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\" id:\"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\" pid:4064 exit_status:137 exited_at:{seconds:1747322079 nanos:289872236}" May 15 15:14:39.337584 containerd[1549]: time="2025-05-15T15:14:39.337449622Z" level=info msg="shim disconnected" id=a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f namespace=k8s.io May 15 15:14:39.337584 containerd[1549]: time="2025-05-15T15:14:39.337480178Z" level=warning msg="cleaning up after shim disconnected" id=a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f namespace=k8s.io May 15 15:14:39.337584 containerd[1549]: time="2025-05-15T15:14:39.337487419Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 15 15:14:39.342062 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f-rootfs.mount: Deactivated successfully. May 15 15:14:39.372789 containerd[1549]: time="2025-05-15T15:14:39.372739498Z" level=info msg="received exit event sandbox_id:\"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\" exit_status:137 exited_at:{seconds:1747322079 nanos:289872236}" May 15 15:14:39.379121 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f-shm.mount: Deactivated successfully. May 15 15:14:39.459364 systemd-networkd[1455]: calidacd95855ff: Link DOWN May 15 15:14:39.459813 systemd-networkd[1455]: calidacd95855ff: Lost carrier May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.457 [INFO][5633] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.457 [INFO][5633] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" iface="eth0" netns="/var/run/netns/cni-c66b3922-5051-1070-c3a0-7d7f940ad52f" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.458 [INFO][5633] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" iface="eth0" netns="/var/run/netns/cni-c66b3922-5051-1070-c3a0-7d7f940ad52f" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.469 [INFO][5633] cni-plugin/dataplane_linux.go 604: Deleted device in netns. ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" after=12.344866ms iface="eth0" netns="/var/run/netns/cni-c66b3922-5051-1070-c3a0-7d7f940ad52f" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.470 [INFO][5633] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.470 [INFO][5633] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.522 [INFO][5643] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" HandleID="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.523 [INFO][5643] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.523 [INFO][5643] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.592 [INFO][5643] ipam/ipam_plugin.go 431: Released address using handleID ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" HandleID="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.592 [INFO][5643] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" HandleID="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.598 [INFO][5643] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:39.602976 containerd[1549]: 2025-05-15 15:14:39.599 [INFO][5633] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" May 15 15:14:39.604330 containerd[1549]: time="2025-05-15T15:14:39.604164168Z" level=info msg="TearDown network for sandbox \"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\" successfully" May 15 15:14:39.604330 containerd[1549]: time="2025-05-15T15:14:39.604207234Z" level=info msg="StopPodSandbox for \"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\" returns successfully" May 15 15:14:39.610524 systemd[1]: run-netns-cni\x2dc66b3922\x2d5051\x2d1070\x2dc3a0\x2d7d7f940ad52f.mount: Deactivated successfully. May 15 15:14:39.757454 kubelet[2694]: I0515 15:14:39.757326 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/826ead8b-3406-4df0-8588-b2656fcca004-tigera-ca-bundle\") pod \"826ead8b-3406-4df0-8588-b2656fcca004\" (UID: \"826ead8b-3406-4df0-8588-b2656fcca004\") " May 15 15:14:39.757454 kubelet[2694]: I0515 15:14:39.757389 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrs64\" (UniqueName: \"kubernetes.io/projected/826ead8b-3406-4df0-8588-b2656fcca004-kube-api-access-xrs64\") pod \"826ead8b-3406-4df0-8588-b2656fcca004\" (UID: \"826ead8b-3406-4df0-8588-b2656fcca004\") " May 15 15:14:39.765302 kubelet[2694]: I0515 15:14:39.765233 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/826ead8b-3406-4df0-8588-b2656fcca004-kube-api-access-xrs64" (OuterVolumeSpecName: "kube-api-access-xrs64") pod "826ead8b-3406-4df0-8588-b2656fcca004" (UID: "826ead8b-3406-4df0-8588-b2656fcca004"). InnerVolumeSpecName "kube-api-access-xrs64". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 15 15:14:39.767195 systemd[1]: var-lib-kubelet-pods-826ead8b\x2d3406\x2d4df0\x2d8588\x2db2656fcca004-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dxrs64.mount: Deactivated successfully. May 15 15:14:39.780470 kubelet[2694]: I0515 15:14:39.780195 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826ead8b-3406-4df0-8588-b2656fcca004-tigera-ca-bundle" (OuterVolumeSpecName: "tigera-ca-bundle") pod "826ead8b-3406-4df0-8588-b2656fcca004" (UID: "826ead8b-3406-4df0-8588-b2656fcca004"). InnerVolumeSpecName "tigera-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 15 15:14:39.858121 kubelet[2694]: I0515 15:14:39.858028 2694 reconciler_common.go:299] "Volume detached for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/826ead8b-3406-4df0-8588-b2656fcca004-tigera-ca-bundle\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:39.858121 kubelet[2694]: I0515 15:14:39.858089 2694 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xrs64\" (UniqueName: \"kubernetes.io/projected/826ead8b-3406-4df0-8588-b2656fcca004-kube-api-access-xrs64\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:40.158668 kubelet[2694]: I0515 15:14:40.158219 2694 scope.go:117] "RemoveContainer" containerID="6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11" May 15 15:14:40.164190 containerd[1549]: time="2025-05-15T15:14:40.163350445Z" level=info msg="RemoveContainer for \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\"" May 15 15:14:40.173547 systemd[1]: Removed slice kubepods-besteffort-pod826ead8b_3406_4df0_8588_b2656fcca004.slice - libcontainer container kubepods-besteffort-pod826ead8b_3406_4df0_8588_b2656fcca004.slice. May 15 15:14:40.173662 systemd[1]: kubepods-besteffort-pod826ead8b_3406_4df0_8588_b2656fcca004.slice: Consumed 344ms CPU time, 16.5M memory peak, 2.2M read from disk, 20K written to disk. May 15 15:14:40.179755 containerd[1549]: time="2025-05-15T15:14:40.179712846Z" level=info msg="RemoveContainer for \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" returns successfully" May 15 15:14:40.181087 kubelet[2694]: I0515 15:14:40.181054 2694 scope.go:117] "RemoveContainer" containerID="6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11" May 15 15:14:40.181405 containerd[1549]: time="2025-05-15T15:14:40.181359001Z" level=error msg="ContainerStatus for \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\": not found" May 15 15:14:40.187022 kubelet[2694]: E0515 15:14:40.186974 2694 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\": not found" containerID="6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11" May 15 15:14:40.187440 kubelet[2694]: I0515 15:14:40.187031 2694 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11"} err="failed to get container status \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\": rpc error: code = NotFound desc = an error occurred when try to find container \"6618b7a29ffb8b688d7e648bb3b5814a54167539726557f089cd93d5b0a57c11\": not found" May 15 15:14:40.226945 kubelet[2694]: I0515 15:14:40.226885 2694 memory_manager.go:355] "RemoveStaleState removing state" podUID="1c80ddf0-e33c-4576-8bd9-03a8cc5f33e7" containerName="calico-apiserver" May 15 15:14:40.226945 kubelet[2694]: I0515 15:14:40.226954 2694 memory_manager.go:355] "RemoveStaleState removing state" podUID="52c8dc92-1a03-4e1d-bc5b-3844d9a477ad" containerName="calico-apiserver" May 15 15:14:40.227139 kubelet[2694]: I0515 15:14:40.226964 2694 memory_manager.go:355] "RemoveStaleState removing state" podUID="826ead8b-3406-4df0-8588-b2656fcca004" containerName="calico-kube-controllers" May 15 15:14:40.237566 systemd[1]: Created slice kubepods-besteffort-podaf3e1df9_cdaa_46cc_b596_17f16d774cdc.slice - libcontainer container kubepods-besteffort-podaf3e1df9_cdaa_46cc_b596_17f16d774cdc.slice. May 15 15:14:40.255821 systemd[1]: var-lib-kubelet-pods-826ead8b\x2d3406\x2d4df0\x2d8588\x2db2656fcca004-volume\x2dsubpaths-tigera\x2dca\x2dbundle-calico\x2dkube\x2dcontrollers-1.mount: Deactivated successfully. May 15 15:14:40.371802 kubelet[2694]: I0515 15:14:40.371687 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af3e1df9-cdaa-46cc-b596-17f16d774cdc-tigera-ca-bundle\") pod \"calico-kube-controllers-6c6bc659b-l4t4l\" (UID: \"af3e1df9-cdaa-46cc-b596-17f16d774cdc\") " pod="calico-system/calico-kube-controllers-6c6bc659b-l4t4l" May 15 15:14:40.372035 kubelet[2694]: I0515 15:14:40.371821 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snq7k\" (UniqueName: \"kubernetes.io/projected/af3e1df9-cdaa-46cc-b596-17f16d774cdc-kube-api-access-snq7k\") pod \"calico-kube-controllers-6c6bc659b-l4t4l\" (UID: \"af3e1df9-cdaa-46cc-b596-17f16d774cdc\") " pod="calico-system/calico-kube-controllers-6c6bc659b-l4t4l" May 15 15:14:40.486607 kubelet[2694]: I0515 15:14:40.486211 2694 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="826ead8b-3406-4df0-8588-b2656fcca004" path="/var/lib/kubelet/pods/826ead8b-3406-4df0-8588-b2656fcca004/volumes" May 15 15:14:40.544369 containerd[1549]: time="2025-05-15T15:14:40.544309211Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6c6bc659b-l4t4l,Uid:af3e1df9-cdaa-46cc-b596-17f16d774cdc,Namespace:calico-system,Attempt:0,}" May 15 15:14:40.704806 systemd-networkd[1455]: cali301a2bd2afc: Link UP May 15 15:14:40.706321 systemd-networkd[1455]: cali301a2bd2afc: Gained carrier May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.594 [INFO][5664] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0 calico-kube-controllers-6c6bc659b- calico-system af3e1df9-cdaa-46cc-b596-17f16d774cdc 1277 0 2025-05-15 15:14:40 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6c6bc659b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4334.0.0-a-073c24674a calico-kube-controllers-6c6bc659b-l4t4l eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali301a2bd2afc [] []}} ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Namespace="calico-system" Pod="calico-kube-controllers-6c6bc659b-l4t4l" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.595 [INFO][5664] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Namespace="calico-system" Pod="calico-kube-controllers-6c6bc659b-l4t4l" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.630 [INFO][5675] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" HandleID="k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.642 [INFO][5675] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" HandleID="k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00011bc40), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4334.0.0-a-073c24674a", "pod":"calico-kube-controllers-6c6bc659b-l4t4l", "timestamp":"2025-05-15 15:14:40.630498977 +0000 UTC"}, Hostname:"ci-4334.0.0-a-073c24674a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.642 [INFO][5675] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.642 [INFO][5675] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.642 [INFO][5675] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-073c24674a' May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.645 [INFO][5675] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.654 [INFO][5675] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.660 [INFO][5675] ipam/ipam.go 489: Trying affinity for 192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.664 [INFO][5675] ipam/ipam.go 155: Attempting to load block cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.668 [INFO][5675] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.50.64/26 host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.668 [INFO][5675] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.50.64/26 handle="k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.671 [INFO][5675] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.677 [INFO][5675] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.50.64/26 handle="k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.693 [INFO][5675] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.50.73/26] block=192.168.50.64/26 handle="k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.693 [INFO][5675] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.50.73/26] handle="k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" host="ci-4334.0.0-a-073c24674a" May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.694 [INFO][5675] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:14:40.732232 containerd[1549]: 2025-05-15 15:14:40.694 [INFO][5675] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.50.73/26] IPv6=[] ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" HandleID="k8s-pod-network.02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" May 15 15:14:40.735456 containerd[1549]: 2025-05-15 15:14:40.698 [INFO][5664] cni-plugin/k8s.go 386: Populated endpoint ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Namespace="calico-system" Pod="calico-kube-controllers-6c6bc659b-l4t4l" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0", GenerateName:"calico-kube-controllers-6c6bc659b-", Namespace:"calico-system", SelfLink:"", UID:"af3e1df9-cdaa-46cc-b596-17f16d774cdc", ResourceVersion:"1277", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 14, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6c6bc659b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"", Pod:"calico-kube-controllers-6c6bc659b-l4t4l", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.50.73/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali301a2bd2afc", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:40.735456 containerd[1549]: 2025-05-15 15:14:40.699 [INFO][5664] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.50.73/32] ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Namespace="calico-system" Pod="calico-kube-controllers-6c6bc659b-l4t4l" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" May 15 15:14:40.735456 containerd[1549]: 2025-05-15 15:14:40.699 [INFO][5664] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali301a2bd2afc ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Namespace="calico-system" Pod="calico-kube-controllers-6c6bc659b-l4t4l" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" May 15 15:14:40.735456 containerd[1549]: 2025-05-15 15:14:40.706 [INFO][5664] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Namespace="calico-system" Pod="calico-kube-controllers-6c6bc659b-l4t4l" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" May 15 15:14:40.735456 containerd[1549]: 2025-05-15 15:14:40.707 [INFO][5664] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Namespace="calico-system" Pod="calico-kube-controllers-6c6bc659b-l4t4l" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0", GenerateName:"calico-kube-controllers-6c6bc659b-", Namespace:"calico-system", SelfLink:"", UID:"af3e1df9-cdaa-46cc-b596-17f16d774cdc", ResourceVersion:"1277", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 15, 14, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6c6bc659b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-073c24674a", ContainerID:"02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a", Pod:"calico-kube-controllers-6c6bc659b-l4t4l", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.50.73/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali301a2bd2afc", MAC:"f2:9a:6c:14:b7:b6", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 15:14:40.735456 containerd[1549]: 2025-05-15 15:14:40.728 [INFO][5664] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" Namespace="calico-system" Pod="calico-kube-controllers-6c6bc659b-l4t4l" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--6c6bc659b--l4t4l-eth0" May 15 15:14:40.774734 containerd[1549]: time="2025-05-15T15:14:40.774564861Z" level=info msg="connecting to shim 02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a" address="unix:///run/containerd/s/898c698d65e3a7dd2b6e470a5cb10e71d68782715fcdbcf6117a462ca8ee0c30" namespace=k8s.io protocol=ttrpc version=3 May 15 15:14:40.825716 systemd[1]: Started cri-containerd-02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a.scope - libcontainer container 02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a. May 15 15:14:40.905052 containerd[1549]: time="2025-05-15T15:14:40.904974545Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6c6bc659b-l4t4l,Uid:af3e1df9-cdaa-46cc-b596-17f16d774cdc,Namespace:calico-system,Attempt:0,} returns sandbox id \"02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a\"" May 15 15:14:40.935014 containerd[1549]: time="2025-05-15T15:14:40.934577956Z" level=info msg="CreateContainer within sandbox \"02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" May 15 15:14:40.939300 containerd[1549]: time="2025-05-15T15:14:40.939257987Z" level=info msg="Container 331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756: CDI devices from CRI Config.CDIDevices: []" May 15 15:14:40.945913 containerd[1549]: time="2025-05-15T15:14:40.945759439Z" level=info msg="CreateContainer within sandbox \"02e6b228a78f8c9f28892f0e9246f66ddfcce5acc826b0cb14e88e11fd0b4c9a\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756\"" May 15 15:14:40.946916 containerd[1549]: time="2025-05-15T15:14:40.946857318Z" level=info msg="StartContainer for \"331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756\"" May 15 15:14:40.950235 containerd[1549]: time="2025-05-15T15:14:40.950052766Z" level=info msg="connecting to shim 331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756" address="unix:///run/containerd/s/898c698d65e3a7dd2b6e470a5cb10e71d68782715fcdbcf6117a462ca8ee0c30" protocol=ttrpc version=3 May 15 15:14:40.985203 systemd[1]: Started cri-containerd-331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756.scope - libcontainer container 331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756. May 15 15:14:41.047673 containerd[1549]: time="2025-05-15T15:14:41.046310538Z" level=info msg="StartContainer for \"331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756\" returns successfully" May 15 15:14:41.282568 containerd[1549]: time="2025-05-15T15:14:41.282335208Z" level=info msg="TaskExit event in podsandbox handler container_id:\"331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756\" id:\"6bcce80fe87327c51ce3141093afe9b9ff0aa2c3df995e5377f88496a8ef7446\" pid:5792 exited_at:{seconds:1747322081 nanos:281886384}" May 15 15:14:41.300066 kubelet[2694]: I0515 15:14:41.299808 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-6c6bc659b-l4t4l" podStartSLOduration=1.2997877789999999 podStartE2EDuration="1.299787779s" podCreationTimestamp="2025-05-15 15:14:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:14:41.185998903 +0000 UTC m=+74.866871929" watchObservedRunningTime="2025-05-15 15:14:41.299787779 +0000 UTC m=+74.980660806" May 15 15:14:41.489570 kubelet[2694]: E0515 15:14:41.489444 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:14:41.743302 systemd-networkd[1455]: cali301a2bd2afc: Gained IPv6LL May 15 15:14:43.306921 systemd[1]: Started sshd@17-24.144.90.91:22-139.178.68.195:48072.service - OpenSSH per-connection server daemon (139.178.68.195:48072). May 15 15:14:43.418021 sshd[5839]: Accepted publickey for core from 139.178.68.195 port 48072 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:43.421480 sshd-session[5839]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:43.429038 systemd-logind[1527]: New session 15 of user core. May 15 15:14:43.436287 systemd[1]: Started session-15.scope - Session 15 of User core. May 15 15:14:43.436559 systemd[1]: cri-containerd-c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5.scope: Deactivated successfully. May 15 15:14:43.437582 systemd[1]: cri-containerd-c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5.scope: Consumed 394ms CPU time, 41.1M memory peak, 28M read from disk. May 15 15:14:43.440483 containerd[1549]: time="2025-05-15T15:14:43.440433527Z" level=info msg="received exit event container_id:\"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" id:\"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" pid:3250 exit_status:1 exited_at:{seconds:1747322083 nanos:439393114}" May 15 15:14:43.442791 containerd[1549]: time="2025-05-15T15:14:43.440853403Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" id:\"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" pid:3250 exit_status:1 exited_at:{seconds:1747322083 nanos:439393114}" May 15 15:14:43.484690 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5-rootfs.mount: Deactivated successfully. May 15 15:14:43.498347 containerd[1549]: time="2025-05-15T15:14:43.498218794Z" level=info msg="StopContainer for \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" returns successfully" May 15 15:14:43.499473 containerd[1549]: time="2025-05-15T15:14:43.499430253Z" level=info msg="StopPodSandbox for \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\"" May 15 15:14:43.499660 containerd[1549]: time="2025-05-15T15:14:43.499637367Z" level=info msg="Container to stop \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 15 15:14:43.511208 systemd[1]: cri-containerd-4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58.scope: Deactivated successfully. May 15 15:14:43.514781 containerd[1549]: time="2025-05-15T15:14:43.514649084Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" id:\"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" pid:3126 exit_status:137 exited_at:{seconds:1747322083 nanos:514213313}" May 15 15:14:43.569848 containerd[1549]: time="2025-05-15T15:14:43.569089178Z" level=info msg="shim disconnected" id=4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58 namespace=k8s.io May 15 15:14:43.569848 containerd[1549]: time="2025-05-15T15:14:43.569136065Z" level=warning msg="cleaning up after shim disconnected" id=4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58 namespace=k8s.io May 15 15:14:43.569848 containerd[1549]: time="2025-05-15T15:14:43.569149896Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 15 15:14:43.578086 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58-rootfs.mount: Deactivated successfully. May 15 15:14:43.615914 containerd[1549]: time="2025-05-15T15:14:43.615783822Z" level=info msg="received exit event sandbox_id:\"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" exit_status:137 exited_at:{seconds:1747322083 nanos:514213313}" May 15 15:14:43.617110 containerd[1549]: time="2025-05-15T15:14:43.616572354Z" level=info msg="TearDown network for sandbox \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" successfully" May 15 15:14:43.617110 containerd[1549]: time="2025-05-15T15:14:43.616605459Z" level=info msg="StopPodSandbox for \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" returns successfully" May 15 15:14:43.623896 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58-shm.mount: Deactivated successfully. May 15 15:14:43.713188 sshd[5847]: Connection closed by 139.178.68.195 port 48072 May 15 15:14:43.716261 sshd-session[5839]: pam_unix(sshd:session): session closed for user core May 15 15:14:43.722457 systemd[1]: sshd@17-24.144.90.91:22-139.178.68.195:48072.service: Deactivated successfully. May 15 15:14:43.725164 systemd[1]: session-15.scope: Deactivated successfully. May 15 15:14:43.726674 systemd-logind[1527]: Session 15 logged out. Waiting for processes to exit. May 15 15:14:43.729263 systemd-logind[1527]: Removed session 15. May 15 15:14:43.797151 kubelet[2694]: I0515 15:14:43.797099 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12abe38c-3744-42fb-bd9c-1466a933b3e0-tigera-ca-bundle\") pod \"12abe38c-3744-42fb-bd9c-1466a933b3e0\" (UID: \"12abe38c-3744-42fb-bd9c-1466a933b3e0\") " May 15 15:14:43.797151 kubelet[2694]: I0515 15:14:43.797153 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/12abe38c-3744-42fb-bd9c-1466a933b3e0-typha-certs\") pod \"12abe38c-3744-42fb-bd9c-1466a933b3e0\" (UID: \"12abe38c-3744-42fb-bd9c-1466a933b3e0\") " May 15 15:14:43.798028 kubelet[2694]: I0515 15:14:43.797184 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkwfq\" (UniqueName: \"kubernetes.io/projected/12abe38c-3744-42fb-bd9c-1466a933b3e0-kube-api-access-kkwfq\") pod \"12abe38c-3744-42fb-bd9c-1466a933b3e0\" (UID: \"12abe38c-3744-42fb-bd9c-1466a933b3e0\") " May 15 15:14:43.806956 kubelet[2694]: I0515 15:14:43.804528 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12abe38c-3744-42fb-bd9c-1466a933b3e0-kube-api-access-kkwfq" (OuterVolumeSpecName: "kube-api-access-kkwfq") pod "12abe38c-3744-42fb-bd9c-1466a933b3e0" (UID: "12abe38c-3744-42fb-bd9c-1466a933b3e0"). InnerVolumeSpecName "kube-api-access-kkwfq". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 15 15:14:43.806381 systemd[1]: var-lib-kubelet-pods-12abe38c\x2d3744\x2d42fb\x2dbd9c\x2d1466a933b3e0-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkkwfq.mount: Deactivated successfully. May 15 15:14:43.810599 systemd[1]: var-lib-kubelet-pods-12abe38c\x2d3744\x2d42fb\x2dbd9c\x2d1466a933b3e0-volumes-kubernetes.io\x7esecret-typha\x2dcerts.mount: Deactivated successfully. May 15 15:14:43.811985 kubelet[2694]: I0515 15:14:43.810797 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12abe38c-3744-42fb-bd9c-1466a933b3e0-typha-certs" (OuterVolumeSpecName: "typha-certs") pod "12abe38c-3744-42fb-bd9c-1466a933b3e0" (UID: "12abe38c-3744-42fb-bd9c-1466a933b3e0"). InnerVolumeSpecName "typha-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 15 15:14:43.813002 kubelet[2694]: I0515 15:14:43.812947 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12abe38c-3744-42fb-bd9c-1466a933b3e0-tigera-ca-bundle" (OuterVolumeSpecName: "tigera-ca-bundle") pod "12abe38c-3744-42fb-bd9c-1466a933b3e0" (UID: "12abe38c-3744-42fb-bd9c-1466a933b3e0"). InnerVolumeSpecName "tigera-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 15 15:14:43.898298 kubelet[2694]: I0515 15:14:43.898138 2694 reconciler_common.go:299] "Volume detached for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12abe38c-3744-42fb-bd9c-1466a933b3e0-tigera-ca-bundle\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:43.898298 kubelet[2694]: I0515 15:14:43.898184 2694 reconciler_common.go:299] "Volume detached for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/12abe38c-3744-42fb-bd9c-1466a933b3e0-typha-certs\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:43.898298 kubelet[2694]: I0515 15:14:43.898199 2694 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kkwfq\" (UniqueName: \"kubernetes.io/projected/12abe38c-3744-42fb-bd9c-1466a933b3e0-kube-api-access-kkwfq\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:14:44.175573 kubelet[2694]: I0515 15:14:44.175373 2694 scope.go:117] "RemoveContainer" containerID="c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5" May 15 15:14:44.186466 systemd[1]: Removed slice kubepods-besteffort-pod12abe38c_3744_42fb_bd9c_1466a933b3e0.slice - libcontainer container kubepods-besteffort-pod12abe38c_3744_42fb_bd9c_1466a933b3e0.slice. May 15 15:14:44.186792 systemd[1]: kubepods-besteffort-pod12abe38c_3744_42fb_bd9c_1466a933b3e0.slice: Consumed 435ms CPU time, 41.4M memory peak, 28M read from disk. May 15 15:14:44.189084 containerd[1549]: time="2025-05-15T15:14:44.187631718Z" level=info msg="RemoveContainer for \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\"" May 15 15:14:44.194273 containerd[1549]: time="2025-05-15T15:14:44.194183655Z" level=info msg="RemoveContainer for \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" returns successfully" May 15 15:14:44.196155 kubelet[2694]: I0515 15:14:44.196116 2694 scope.go:117] "RemoveContainer" containerID="c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5" May 15 15:14:44.196705 kubelet[2694]: E0515 15:14:44.196542 2694 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\": not found" containerID="c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5" May 15 15:14:44.196705 kubelet[2694]: I0515 15:14:44.196568 2694 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5"} err="failed to get container status \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\": rpc error: code = NotFound desc = an error occurred when try to find container \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\": not found" May 15 15:14:44.196823 containerd[1549]: time="2025-05-15T15:14:44.196381392Z" level=error msg="ContainerStatus for \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c46b1154d1acbe3a4fa4b39ec5c0167119b53bcee34060716278ce037e76a6e5\": not found" May 15 15:14:44.481348 systemd[1]: var-lib-kubelet-pods-12abe38c\x2d3744\x2d42fb\x2dbd9c\x2d1466a933b3e0-volume\x2dsubpaths-tigera\x2dca\x2dbundle-calico\x2dtypha-1.mount: Deactivated successfully. May 15 15:14:44.485833 kubelet[2694]: I0515 15:14:44.485791 2694 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12abe38c-3744-42fb-bd9c-1466a933b3e0" path="/var/lib/kubelet/pods/12abe38c-3744-42fb-bd9c-1466a933b3e0/volumes" May 15 15:14:48.730178 systemd[1]: Started sshd@18-24.144.90.91:22-139.178.68.195:44400.service - OpenSSH per-connection server daemon (139.178.68.195:44400). May 15 15:14:48.788470 sshd[6013]: Accepted publickey for core from 139.178.68.195 port 44400 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:48.791072 sshd-session[6013]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:48.797187 systemd-logind[1527]: New session 16 of user core. May 15 15:14:48.808211 systemd[1]: Started session-16.scope - Session 16 of User core. May 15 15:14:48.946645 sshd[6015]: Connection closed by 139.178.68.195 port 44400 May 15 15:14:48.947456 sshd-session[6013]: pam_unix(sshd:session): session closed for user core May 15 15:14:48.963628 systemd[1]: sshd@18-24.144.90.91:22-139.178.68.195:44400.service: Deactivated successfully. May 15 15:14:48.966311 systemd[1]: session-16.scope: Deactivated successfully. May 15 15:14:48.967490 systemd-logind[1527]: Session 16 logged out. Waiting for processes to exit. May 15 15:14:48.973166 systemd[1]: Started sshd@19-24.144.90.91:22-139.178.68.195:44402.service - OpenSSH per-connection server daemon (139.178.68.195:44402). May 15 15:14:48.974701 systemd-logind[1527]: Removed session 16. May 15 15:14:49.029610 sshd[6027]: Accepted publickey for core from 139.178.68.195 port 44402 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:49.031268 sshd-session[6027]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:49.037087 systemd-logind[1527]: New session 17 of user core. May 15 15:14:49.041177 systemd[1]: Started session-17.scope - Session 17 of User core. May 15 15:14:49.446186 sshd[6029]: Connection closed by 139.178.68.195 port 44402 May 15 15:14:49.446823 sshd-session[6027]: pam_unix(sshd:session): session closed for user core May 15 15:14:49.458007 systemd[1]: sshd@19-24.144.90.91:22-139.178.68.195:44402.service: Deactivated successfully. May 15 15:14:49.460654 systemd[1]: session-17.scope: Deactivated successfully. May 15 15:14:49.464010 systemd-logind[1527]: Session 17 logged out. Waiting for processes to exit. May 15 15:14:49.469225 systemd[1]: Started sshd@20-24.144.90.91:22-139.178.68.195:44408.service - OpenSSH per-connection server daemon (139.178.68.195:44408). May 15 15:14:49.472666 systemd-logind[1527]: Removed session 17. May 15 15:14:49.525032 sshd[6039]: Accepted publickey for core from 139.178.68.195 port 44408 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:49.527208 sshd-session[6039]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:49.532590 systemd-logind[1527]: New session 18 of user core. May 15 15:14:49.537157 systemd[1]: Started session-18.scope - Session 18 of User core. May 15 15:14:50.660664 sshd[6041]: Connection closed by 139.178.68.195 port 44408 May 15 15:14:50.662111 sshd-session[6039]: pam_unix(sshd:session): session closed for user core May 15 15:14:50.673257 systemd[1]: sshd@20-24.144.90.91:22-139.178.68.195:44408.service: Deactivated successfully. May 15 15:14:50.678558 systemd[1]: session-18.scope: Deactivated successfully. May 15 15:14:50.680865 systemd-logind[1527]: Session 18 logged out. Waiting for processes to exit. May 15 15:14:50.687230 systemd-logind[1527]: Removed session 18. May 15 15:14:50.691920 systemd[1]: Started sshd@21-24.144.90.91:22-139.178.68.195:44418.service - OpenSSH per-connection server daemon (139.178.68.195:44418). May 15 15:14:50.803989 sshd[6079]: Accepted publickey for core from 139.178.68.195 port 44418 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:50.806054 sshd-session[6079]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:50.814340 systemd-logind[1527]: New session 19 of user core. May 15 15:14:50.821220 systemd[1]: Started session-19.scope - Session 19 of User core. May 15 15:14:51.193629 sshd[6090]: Connection closed by 139.178.68.195 port 44418 May 15 15:14:51.194380 sshd-session[6079]: pam_unix(sshd:session): session closed for user core May 15 15:14:51.210166 systemd[1]: sshd@21-24.144.90.91:22-139.178.68.195:44418.service: Deactivated successfully. May 15 15:14:51.212826 systemd[1]: session-19.scope: Deactivated successfully. May 15 15:14:51.214388 systemd-logind[1527]: Session 19 logged out. Waiting for processes to exit. May 15 15:14:51.222065 systemd[1]: Started sshd@22-24.144.90.91:22-139.178.68.195:44432.service - OpenSSH per-connection server daemon (139.178.68.195:44432). May 15 15:14:51.226479 systemd-logind[1527]: Removed session 19. May 15 15:14:51.303402 sshd[6114]: Accepted publickey for core from 139.178.68.195 port 44432 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:51.305017 sshd-session[6114]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:51.311311 systemd-logind[1527]: New session 20 of user core. May 15 15:14:51.321220 systemd[1]: Started session-20.scope - Session 20 of User core. May 15 15:14:51.470153 sshd[6116]: Connection closed by 139.178.68.195 port 44432 May 15 15:14:51.469984 sshd-session[6114]: pam_unix(sshd:session): session closed for user core May 15 15:14:51.477404 systemd-logind[1527]: Session 20 logged out. Waiting for processes to exit. May 15 15:14:51.478293 systemd[1]: sshd@22-24.144.90.91:22-139.178.68.195:44432.service: Deactivated successfully. May 15 15:14:51.480720 systemd[1]: session-20.scope: Deactivated successfully. May 15 15:14:51.483573 systemd-logind[1527]: Removed session 20. May 15 15:14:56.488073 systemd[1]: Started sshd@23-24.144.90.91:22-139.178.68.195:35476.service - OpenSSH per-connection server daemon (139.178.68.195:35476). May 15 15:14:56.608051 sshd[6227]: Accepted publickey for core from 139.178.68.195 port 35476 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:14:56.610049 sshd-session[6227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:14:56.615340 systemd-logind[1527]: New session 21 of user core. May 15 15:14:56.622711 systemd[1]: Started session-21.scope - Session 21 of User core. May 15 15:14:56.831977 sshd[6229]: Connection closed by 139.178.68.195 port 35476 May 15 15:14:56.833252 sshd-session[6227]: pam_unix(sshd:session): session closed for user core May 15 15:14:56.838264 systemd[1]: sshd@23-24.144.90.91:22-139.178.68.195:35476.service: Deactivated successfully. May 15 15:14:56.838513 systemd-logind[1527]: Session 21 logged out. Waiting for processes to exit. May 15 15:14:56.841420 systemd[1]: session-21.scope: Deactivated successfully. May 15 15:14:56.844626 systemd-logind[1527]: Removed session 21. May 15 15:14:57.482598 kubelet[2694]: E0515 15:14:57.482546 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:01.483019 kubelet[2694]: E0515 15:15:01.482511 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:01.483019 kubelet[2694]: E0515 15:15:01.482514 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:01.854351 systemd[1]: Started sshd@24-24.144.90.91:22-139.178.68.195:35490.service - OpenSSH per-connection server daemon (139.178.68.195:35490). May 15 15:15:01.884690 containerd[1549]: time="2025-05-15T15:15:01.884628636Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" id:\"5cb7cb8da82f36ffba16c16e6836a5afdfc12005434a46d09dbdfe8d685f8a91\" pid:6352 exit_status:1 exited_at:{seconds:1747322101 nanos:882453266}" May 15 15:15:01.935847 sshd[6365]: Accepted publickey for core from 139.178.68.195 port 35490 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:15:01.938297 sshd-session[6365]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:15:01.945256 systemd-logind[1527]: New session 22 of user core. May 15 15:15:01.953319 systemd[1]: Started session-22.scope - Session 22 of User core. May 15 15:15:02.104419 sshd[6368]: Connection closed by 139.178.68.195 port 35490 May 15 15:15:02.105496 sshd-session[6365]: pam_unix(sshd:session): session closed for user core May 15 15:15:02.110129 systemd[1]: sshd@24-24.144.90.91:22-139.178.68.195:35490.service: Deactivated successfully. May 15 15:15:02.113419 systemd[1]: session-22.scope: Deactivated successfully. May 15 15:15:02.115379 systemd-logind[1527]: Session 22 logged out. Waiting for processes to exit. May 15 15:15:02.118971 systemd-logind[1527]: Removed session 22. May 15 15:15:07.125620 systemd[1]: Started sshd@25-24.144.90.91:22-139.178.68.195:35918.service - OpenSSH per-connection server daemon (139.178.68.195:35918). May 15 15:15:07.237620 sshd[6475]: Accepted publickey for core from 139.178.68.195 port 35918 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:15:07.240111 sshd-session[6475]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:15:07.249070 systemd-logind[1527]: New session 23 of user core. May 15 15:15:07.256266 systemd[1]: Started session-23.scope - Session 23 of User core. May 15 15:15:07.545020 sshd[6477]: Connection closed by 139.178.68.195 port 35918 May 15 15:15:07.544857 sshd-session[6475]: pam_unix(sshd:session): session closed for user core May 15 15:15:07.551013 systemd[1]: sshd@25-24.144.90.91:22-139.178.68.195:35918.service: Deactivated successfully. May 15 15:15:07.553697 systemd[1]: session-23.scope: Deactivated successfully. May 15 15:15:07.555401 systemd-logind[1527]: Session 23 logged out. Waiting for processes to exit. May 15 15:15:07.558315 systemd-logind[1527]: Removed session 23. May 15 15:15:11.224097 containerd[1549]: time="2025-05-15T15:15:11.224045721Z" level=info msg="TaskExit event in podsandbox handler container_id:\"331e7f4c73b06f986cf264c0ba7696cab475eb9d69715388965ead898224a756\" id:\"262a60ee5e63bd0fa22e7fb00d6adc41f7265c43eacdf7b47f79791121c8f3ba\" pid:6592 exited_at:{seconds:1747322111 nanos:223212443}" May 15 15:15:12.562966 systemd[1]: Started sshd@26-24.144.90.91:22-139.178.68.195:35934.service - OpenSSH per-connection server daemon (139.178.68.195:35934). May 15 15:15:12.658484 sshd[6626]: Accepted publickey for core from 139.178.68.195 port 35934 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:15:12.660201 sshd-session[6626]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:15:12.669001 systemd-logind[1527]: New session 24 of user core. May 15 15:15:12.676488 systemd[1]: Started session-24.scope - Session 24 of User core. May 15 15:15:12.966877 sshd[6629]: Connection closed by 139.178.68.195 port 35934 May 15 15:15:12.969134 sshd-session[6626]: pam_unix(sshd:session): session closed for user core May 15 15:15:12.976601 systemd[1]: sshd@26-24.144.90.91:22-139.178.68.195:35934.service: Deactivated successfully. May 15 15:15:12.978892 systemd[1]: session-24.scope: Deactivated successfully. May 15 15:15:12.980626 systemd-logind[1527]: Session 24 logged out. Waiting for processes to exit. May 15 15:15:12.983502 systemd-logind[1527]: Removed session 24. May 15 15:15:14.482327 kubelet[2694]: E0515 15:15:14.482220 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:14.483572 kubelet[2694]: E0515 15:15:14.483539 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:17.989907 systemd[1]: Started sshd@27-24.144.90.91:22-139.178.68.195:54144.service - OpenSSH per-connection server daemon (139.178.68.195:54144). May 15 15:15:18.108887 sshd[6733]: Accepted publickey for core from 139.178.68.195 port 54144 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:15:18.112371 sshd-session[6733]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:15:18.123352 systemd-logind[1527]: New session 25 of user core. May 15 15:15:18.127247 systemd[1]: Started session-25.scope - Session 25 of User core. May 15 15:15:18.349347 sshd[6735]: Connection closed by 139.178.68.195 port 54144 May 15 15:15:18.350076 sshd-session[6733]: pam_unix(sshd:session): session closed for user core May 15 15:15:18.356258 systemd[1]: sshd@27-24.144.90.91:22-139.178.68.195:54144.service: Deactivated successfully. May 15 15:15:18.360646 systemd[1]: session-25.scope: Deactivated successfully. May 15 15:15:18.362193 systemd-logind[1527]: Session 25 logged out. Waiting for processes to exit. May 15 15:15:18.364402 systemd-logind[1527]: Removed session 25. May 15 15:15:20.325252 containerd[1549]: time="2025-05-15T15:15:20.325045579Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" id:\"b0d68c12ffe86aa3f864788f6279a9a279378a9abe756fb2dbf8dcb69a71a94e\" pid:6797 exited_at:{seconds:1747322120 nanos:324462669}" May 15 15:15:20.327130 containerd[1549]: time="2025-05-15T15:15:20.327070025Z" level=info msg="StopContainer for \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" with timeout 5 (s)" May 15 15:15:20.328283 containerd[1549]: time="2025-05-15T15:15:20.327414124Z" level=info msg="Stop container \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" with signal terminated" May 15 15:15:20.351208 systemd[1]: cri-containerd-c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e.scope: Deactivated successfully. May 15 15:15:20.351527 systemd[1]: cri-containerd-c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e.scope: Consumed 10.952s CPU time, 203.9M memory peak, 80.9M read from disk, 4M written to disk. May 15 15:15:20.353842 containerd[1549]: time="2025-05-15T15:15:20.353803426Z" level=info msg="received exit event container_id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" pid:3706 exited_at:{seconds:1747322120 nanos:353411251}" May 15 15:15:20.354910 containerd[1549]: time="2025-05-15T15:15:20.354800847Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" id:\"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" pid:3706 exited_at:{seconds:1747322120 nanos:353411251}" May 15 15:15:20.382838 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e-rootfs.mount: Deactivated successfully. May 15 15:15:20.393363 containerd[1549]: time="2025-05-15T15:15:20.393308852Z" level=info msg="StopContainer for \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" returns successfully" May 15 15:15:20.394190 containerd[1549]: time="2025-05-15T15:15:20.394060962Z" level=info msg="StopPodSandbox for \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\"" May 15 15:15:20.394254 containerd[1549]: time="2025-05-15T15:15:20.394227375Z" level=info msg="Container to stop \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 15 15:15:20.394254 containerd[1549]: time="2025-05-15T15:15:20.394245260Z" level=info msg="Container to stop \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 15 15:15:20.394327 containerd[1549]: time="2025-05-15T15:15:20.394253913Z" level=info msg="Container to stop \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 15 15:15:20.403332 systemd[1]: cri-containerd-887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82.scope: Deactivated successfully. May 15 15:15:20.406689 containerd[1549]: time="2025-05-15T15:15:20.406647174Z" level=info msg="TaskExit event in podsandbox handler container_id:\"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" id:\"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" pid:3212 exit_status:137 exited_at:{seconds:1747322120 nanos:405149989}" May 15 15:15:20.438782 containerd[1549]: time="2025-05-15T15:15:20.438556151Z" level=info msg="received exit event sandbox_id:\"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" exit_status:137 exited_at:{seconds:1747322120 nanos:405149989}" May 15 15:15:20.439802 containerd[1549]: time="2025-05-15T15:15:20.439308896Z" level=info msg="shim disconnected" id=887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82 namespace=k8s.io May 15 15:15:20.439802 containerd[1549]: time="2025-05-15T15:15:20.439612029Z" level=warning msg="cleaning up after shim disconnected" id=887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82 namespace=k8s.io May 15 15:15:20.439802 containerd[1549]: time="2025-05-15T15:15:20.439621848Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 15 15:15:20.440024 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82-rootfs.mount: Deactivated successfully. May 15 15:15:20.445091 containerd[1549]: time="2025-05-15T15:15:20.443038844Z" level=info msg="TearDown network for sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" successfully" May 15 15:15:20.445091 containerd[1549]: time="2025-05-15T15:15:20.443068657Z" level=info msg="StopPodSandbox for \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" returns successfully" May 15 15:15:20.446713 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82-shm.mount: Deactivated successfully. May 15 15:15:20.515419 kubelet[2694]: I0515 15:15:20.515363 2694 memory_manager.go:355] "RemoveStaleState removing state" podUID="06424021-71ea-4cef-bbec-94f6e5d79ae3" containerName="calico-node" May 15 15:15:20.515419 kubelet[2694]: I0515 15:15:20.515397 2694 memory_manager.go:355] "RemoveStaleState removing state" podUID="12abe38c-3744-42fb-bd9c-1466a933b3e0" containerName="calico-typha" May 15 15:15:20.527980 systemd[1]: Created slice kubepods-besteffort-pod8088a1d0_b0ec_4178_bc94_29fb5767db31.slice - libcontainer container kubepods-besteffort-pod8088a1d0_b0ec_4178_bc94_29fb5767db31.slice. May 15 15:15:20.557363 kubelet[2694]: I0515 15:15:20.556950 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-net-dir\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557363 kubelet[2694]: I0515 15:15:20.557010 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-xtables-lock\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557363 kubelet[2694]: I0515 15:15:20.557040 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-bin-dir\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557363 kubelet[2694]: I0515 15:15:20.557075 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghrtz\" (UniqueName: \"kubernetes.io/projected/06424021-71ea-4cef-bbec-94f6e5d79ae3-kube-api-access-ghrtz\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557363 kubelet[2694]: I0515 15:15:20.557101 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-flexvol-driver-host\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557363 kubelet[2694]: I0515 15:15:20.557117 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-log-dir\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557690 kubelet[2694]: I0515 15:15:20.557132 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-policysync\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557690 kubelet[2694]: I0515 15:15:20.557149 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-var-lib-calico\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557690 kubelet[2694]: I0515 15:15:20.557173 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06424021-71ea-4cef-bbec-94f6e5d79ae3-tigera-ca-bundle\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557690 kubelet[2694]: I0515 15:15:20.557186 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-lib-modules\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557690 kubelet[2694]: I0515 15:15:20.557203 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-var-run-calico\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557690 kubelet[2694]: I0515 15:15:20.557228 2694 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/06424021-71ea-4cef-bbec-94f6e5d79ae3-node-certs\") pod \"06424021-71ea-4cef-bbec-94f6e5d79ae3\" (UID: \"06424021-71ea-4cef-bbec-94f6e5d79ae3\") " May 15 15:15:20.557844 kubelet[2694]: I0515 15:15:20.557322 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-policysync\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.557844 kubelet[2694]: I0515 15:15:20.557359 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-flexvol-driver-host\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.557844 kubelet[2694]: I0515 15:15:20.557397 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-lib-modules\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.557844 kubelet[2694]: I0515 15:15:20.557430 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-cni-log-dir\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.557844 kubelet[2694]: I0515 15:15:20.557458 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-xtables-lock\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.558021 kubelet[2694]: I0515 15:15:20.557482 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f48ll\" (UniqueName: \"kubernetes.io/projected/8088a1d0-b0ec-4178-bc94-29fb5767db31-kube-api-access-f48ll\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.558021 kubelet[2694]: I0515 15:15:20.557499 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-var-run-calico\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.558021 kubelet[2694]: I0515 15:15:20.557517 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-var-lib-calico\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.558021 kubelet[2694]: I0515 15:15:20.557538 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-cni-bin-dir\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.558021 kubelet[2694]: I0515 15:15:20.557558 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/8088a1d0-b0ec-4178-bc94-29fb5767db31-node-certs\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.558157 kubelet[2694]: I0515 15:15:20.557580 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/8088a1d0-b0ec-4178-bc94-29fb5767db31-cni-net-dir\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.558157 kubelet[2694]: I0515 15:15:20.557609 2694 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8088a1d0-b0ec-4178-bc94-29fb5767db31-tigera-ca-bundle\") pod \"calico-node-ls2dl\" (UID: \"8088a1d0-b0ec-4178-bc94-29fb5767db31\") " pod="calico-system/calico-node-ls2dl" May 15 15:15:20.558157 kubelet[2694]: I0515 15:15:20.557982 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-net-dir" (OuterVolumeSpecName: "cni-net-dir") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "cni-net-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.558157 kubelet[2694]: I0515 15:15:20.558102 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.558157 kubelet[2694]: I0515 15:15:20.558125 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-bin-dir" (OuterVolumeSpecName: "cni-bin-dir") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "cni-bin-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.559842 kubelet[2694]: I0515 15:15:20.559690 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-flexvol-driver-host" (OuterVolumeSpecName: "flexvol-driver-host") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "flexvol-driver-host". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.559842 kubelet[2694]: I0515 15:15:20.559732 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-log-dir" (OuterVolumeSpecName: "cni-log-dir") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "cni-log-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.559842 kubelet[2694]: I0515 15:15:20.559751 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-policysync" (OuterVolumeSpecName: "policysync") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "policysync". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.559842 kubelet[2694]: I0515 15:15:20.559775 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-var-lib-calico" (OuterVolumeSpecName: "var-lib-calico") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "var-lib-calico". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.565247 kubelet[2694]: I0515 15:15:20.564864 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06424021-71ea-4cef-bbec-94f6e5d79ae3-kube-api-access-ghrtz" (OuterVolumeSpecName: "kube-api-access-ghrtz") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "kube-api-access-ghrtz". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 15 15:15:20.565247 kubelet[2694]: I0515 15:15:20.565107 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-var-run-calico" (OuterVolumeSpecName: "var-run-calico") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "var-run-calico". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.565247 kubelet[2694]: I0515 15:15:20.565129 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 15 15:15:20.566569 systemd[1]: var-lib-kubelet-pods-06424021\x2d71ea\x2d4cef\x2dbbec\x2d94f6e5d79ae3-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dghrtz.mount: Deactivated successfully. May 15 15:15:20.570022 kubelet[2694]: I0515 15:15:20.569523 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06424021-71ea-4cef-bbec-94f6e5d79ae3-tigera-ca-bundle" (OuterVolumeSpecName: "tigera-ca-bundle") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "tigera-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 15 15:15:20.571217 systemd[1]: var-lib-kubelet-pods-06424021\x2d71ea\x2d4cef\x2dbbec\x2d94f6e5d79ae3-volume\x2dsubpaths-tigera\x2dca\x2dbundle-calico\x2dnode-1.mount: Deactivated successfully. May 15 15:15:20.573634 kubelet[2694]: I0515 15:15:20.573580 2694 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06424021-71ea-4cef-bbec-94f6e5d79ae3-node-certs" (OuterVolumeSpecName: "node-certs") pod "06424021-71ea-4cef-bbec-94f6e5d79ae3" (UID: "06424021-71ea-4cef-bbec-94f6e5d79ae3"). InnerVolumeSpecName "node-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 15 15:15:20.658754 kubelet[2694]: I0515 15:15:20.658448 2694 reconciler_common.go:299] "Volume detached for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-flexvol-driver-host\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.658754 kubelet[2694]: I0515 15:15:20.658587 2694 reconciler_common.go:299] "Volume detached for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06424021-71ea-4cef-bbec-94f6e5d79ae3-tigera-ca-bundle\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.658754 kubelet[2694]: I0515 15:15:20.658597 2694 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ghrtz\" (UniqueName: \"kubernetes.io/projected/06424021-71ea-4cef-bbec-94f6e5d79ae3-kube-api-access-ghrtz\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.658754 kubelet[2694]: I0515 15:15:20.658610 2694 reconciler_common.go:299] "Volume detached for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-var-lib-calico\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.658754 kubelet[2694]: I0515 15:15:20.658619 2694 reconciler_common.go:299] "Volume detached for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-log-dir\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.658754 kubelet[2694]: I0515 15:15:20.658628 2694 reconciler_common.go:299] "Volume detached for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-policysync\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.658754 kubelet[2694]: I0515 15:15:20.658636 2694 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-lib-modules\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.658754 kubelet[2694]: I0515 15:15:20.658644 2694 reconciler_common.go:299] "Volume detached for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/06424021-71ea-4cef-bbec-94f6e5d79ae3-node-certs\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.660305 kubelet[2694]: I0515 15:15:20.658651 2694 reconciler_common.go:299] "Volume detached for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-var-run-calico\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.660305 kubelet[2694]: I0515 15:15:20.658659 2694 reconciler_common.go:299] "Volume detached for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-net-dir\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.660305 kubelet[2694]: I0515 15:15:20.658666 2694 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-xtables-lock\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.660305 kubelet[2694]: I0515 15:15:20.658675 2694 reconciler_common.go:299] "Volume detached for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/06424021-71ea-4cef-bbec-94f6e5d79ae3-cni-bin-dir\") on node \"ci-4334.0.0-a-073c24674a\" DevicePath \"\"" May 15 15:15:20.834010 kubelet[2694]: E0515 15:15:20.833885 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:20.835148 containerd[1549]: time="2025-05-15T15:15:20.835002728Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ls2dl,Uid:8088a1d0-b0ec-4178-bc94-29fb5767db31,Namespace:calico-system,Attempt:0,}" May 15 15:15:20.852968 containerd[1549]: time="2025-05-15T15:15:20.852739866Z" level=info msg="connecting to shim 0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be" address="unix:///run/containerd/s/7456a8ec0c8096ff76a14ff380fc2b6f2154c56fcfb429960d91b5c911aa7730" namespace=k8s.io protocol=ttrpc version=3 May 15 15:15:20.884226 systemd[1]: Started cri-containerd-0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be.scope - libcontainer container 0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be. May 15 15:15:20.919356 containerd[1549]: time="2025-05-15T15:15:20.919223153Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ls2dl,Uid:8088a1d0-b0ec-4178-bc94-29fb5767db31,Namespace:calico-system,Attempt:0,} returns sandbox id \"0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be\"" May 15 15:15:20.921121 kubelet[2694]: E0515 15:15:20.921046 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:20.925965 containerd[1549]: time="2025-05-15T15:15:20.925900689Z" level=info msg="CreateContainer within sandbox \"0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 15 15:15:20.937699 containerd[1549]: time="2025-05-15T15:15:20.937651887Z" level=info msg="Container e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474: CDI devices from CRI Config.CDIDevices: []" May 15 15:15:20.943768 containerd[1549]: time="2025-05-15T15:15:20.943660714Z" level=info msg="CreateContainer within sandbox \"0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474\"" May 15 15:15:20.946035 containerd[1549]: time="2025-05-15T15:15:20.945703779Z" level=info msg="StartContainer for \"e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474\"" May 15 15:15:20.947519 containerd[1549]: time="2025-05-15T15:15:20.947483139Z" level=info msg="connecting to shim e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474" address="unix:///run/containerd/s/7456a8ec0c8096ff76a14ff380fc2b6f2154c56fcfb429960d91b5c911aa7730" protocol=ttrpc version=3 May 15 15:15:20.972136 systemd[1]: Started cri-containerd-e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474.scope - libcontainer container e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474. May 15 15:15:21.026582 containerd[1549]: time="2025-05-15T15:15:21.026516533Z" level=info msg="StartContainer for \"e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474\" returns successfully" May 15 15:15:21.064296 systemd[1]: cri-containerd-e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474.scope: Deactivated successfully. May 15 15:15:21.064747 systemd[1]: cri-containerd-e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474.scope: Consumed 42ms CPU time, 17.8M memory peak, 9.8M read from disk, 6.3M written to disk. May 15 15:15:21.070099 containerd[1549]: time="2025-05-15T15:15:21.070031482Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474\" id:\"e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474\" pid:6934 exited_at:{seconds:1747322121 nanos:69551871}" May 15 15:15:21.071143 containerd[1549]: time="2025-05-15T15:15:21.071107914Z" level=info msg="received exit event container_id:\"e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474\" id:\"e0c4d24812b3ed3cfc0ec6a15ff1b937b95cde45e4209538a59e92a92762c474\" pid:6934 exited_at:{seconds:1747322121 nanos:69551871}" May 15 15:15:21.297089 kubelet[2694]: E0515 15:15:21.296461 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:21.303695 containerd[1549]: time="2025-05-15T15:15:21.303593094Z" level=info msg="CreateContainer within sandbox \"0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 15 15:15:21.308235 kubelet[2694]: I0515 15:15:21.308185 2694 scope.go:117] "RemoveContainer" containerID="c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e" May 15 15:15:21.318196 containerd[1549]: time="2025-05-15T15:15:21.318148557Z" level=info msg="RemoveContainer for \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\"" May 15 15:15:21.324417 systemd[1]: Removed slice kubepods-besteffort-pod06424021_71ea_4cef_bbec_94f6e5d79ae3.slice - libcontainer container kubepods-besteffort-pod06424021_71ea_4cef_bbec_94f6e5d79ae3.slice. May 15 15:15:21.324542 systemd[1]: kubepods-besteffort-pod06424021_71ea_4cef_bbec_94f6e5d79ae3.slice: Consumed 11.537s CPU time, 224.7M memory peak, 85.3M read from disk, 164.7M written to disk. May 15 15:15:21.357378 containerd[1549]: time="2025-05-15T15:15:21.357308365Z" level=info msg="Container 3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584: CDI devices from CRI Config.CDIDevices: []" May 15 15:15:21.360957 containerd[1549]: time="2025-05-15T15:15:21.360556712Z" level=info msg="RemoveContainer for \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" returns successfully" May 15 15:15:21.361476 kubelet[2694]: I0515 15:15:21.361449 2694 scope.go:117] "RemoveContainer" containerID="642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c" May 15 15:15:21.371782 containerd[1549]: time="2025-05-15T15:15:21.371742685Z" level=info msg="RemoveContainer for \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\"" May 15 15:15:21.382198 containerd[1549]: time="2025-05-15T15:15:21.382035519Z" level=info msg="RemoveContainer for \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\" returns successfully" May 15 15:15:21.382427 kubelet[2694]: I0515 15:15:21.382401 2694 scope.go:117] "RemoveContainer" containerID="5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a" May 15 15:15:21.388090 containerd[1549]: time="2025-05-15T15:15:21.386749947Z" level=info msg="CreateContainer within sandbox \"0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584\"" May 15 15:15:21.387618 systemd[1]: var-lib-kubelet-pods-06424021\x2d71ea\x2d4cef\x2dbbec\x2d94f6e5d79ae3-volumes-kubernetes.io\x7esecret-node\x2dcerts.mount: Deactivated successfully. May 15 15:15:21.395628 containerd[1549]: time="2025-05-15T15:15:21.395256528Z" level=info msg="StartContainer for \"3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584\"" May 15 15:15:21.398554 containerd[1549]: time="2025-05-15T15:15:21.398277357Z" level=info msg="RemoveContainer for \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\"" May 15 15:15:21.399052 containerd[1549]: time="2025-05-15T15:15:21.399024973Z" level=info msg="connecting to shim 3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584" address="unix:///run/containerd/s/7456a8ec0c8096ff76a14ff380fc2b6f2154c56fcfb429960d91b5c911aa7730" protocol=ttrpc version=3 May 15 15:15:21.408712 containerd[1549]: time="2025-05-15T15:15:21.408641220Z" level=info msg="RemoveContainer for \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\" returns successfully" May 15 15:15:21.410416 kubelet[2694]: I0515 15:15:21.410282 2694 scope.go:117] "RemoveContainer" containerID="c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e" May 15 15:15:21.411560 containerd[1549]: time="2025-05-15T15:15:21.411460658Z" level=error msg="ContainerStatus for \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\": not found" May 15 15:15:21.413224 kubelet[2694]: E0515 15:15:21.413181 2694 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\": not found" containerID="c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e" May 15 15:15:21.413398 kubelet[2694]: I0515 15:15:21.413227 2694 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e"} err="failed to get container status \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\": rpc error: code = NotFound desc = an error occurred when try to find container \"c9268ceb1580687a57a8e7864ef312b1b0bd2f3c5c311f276171c489486a907e\": not found" May 15 15:15:21.413398 kubelet[2694]: I0515 15:15:21.413251 2694 scope.go:117] "RemoveContainer" containerID="642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c" May 15 15:15:21.413621 containerd[1549]: time="2025-05-15T15:15:21.413540392Z" level=error msg="ContainerStatus for \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\": not found" May 15 15:15:21.414244 kubelet[2694]: E0515 15:15:21.414191 2694 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\": not found" containerID="642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c" May 15 15:15:21.414487 kubelet[2694]: I0515 15:15:21.414452 2694 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c"} err="failed to get container status \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\": rpc error: code = NotFound desc = an error occurred when try to find container \"642c75b7628d41e263a5b42a5bfd3aa98d6de46cdff6551316d7ab4af5c5947c\": not found" May 15 15:15:21.414576 kubelet[2694]: I0515 15:15:21.414564 2694 scope.go:117] "RemoveContainer" containerID="5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a" May 15 15:15:21.415164 containerd[1549]: time="2025-05-15T15:15:21.415000590Z" level=error msg="ContainerStatus for \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\": not found" May 15 15:15:21.415647 kubelet[2694]: E0515 15:15:21.415561 2694 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\": not found" containerID="5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a" May 15 15:15:21.415647 kubelet[2694]: I0515 15:15:21.415621 2694 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a"} err="failed to get container status \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\": rpc error: code = NotFound desc = an error occurred when try to find container \"5a7eae61a64566ab35f40d3f9076b6983a852bfd988b13f98237e7e820034d9a\": not found" May 15 15:15:21.438519 systemd[1]: Started cri-containerd-3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584.scope - libcontainer container 3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584. May 15 15:15:21.492212 containerd[1549]: time="2025-05-15T15:15:21.492170131Z" level=info msg="StartContainer for \"3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584\" returns successfully" May 15 15:15:22.317027 kubelet[2694]: E0515 15:15:22.316828 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:22.493824 kubelet[2694]: I0515 15:15:22.493774 2694 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06424021-71ea-4cef-bbec-94f6e5d79ae3" path="/var/lib/kubelet/pods/06424021-71ea-4cef-bbec-94f6e5d79ae3/volumes" May 15 15:15:22.510052 systemd[1]: cri-containerd-3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584.scope: Deactivated successfully. May 15 15:15:22.512310 systemd[1]: cri-containerd-3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584.scope: Consumed 924ms CPU time, 267.3M memory peak, 272.9M read from disk. May 15 15:15:22.514310 containerd[1549]: time="2025-05-15T15:15:22.514192799Z" level=info msg="received exit event container_id:\"3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584\" id:\"3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584\" pid:6986 exited_at:{seconds:1747322122 nanos:513543581}" May 15 15:15:22.515866 containerd[1549]: time="2025-05-15T15:15:22.515838098Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584\" id:\"3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584\" pid:6986 exited_at:{seconds:1747322122 nanos:513543581}" May 15 15:15:22.522546 containerd[1549]: time="2025-05-15T15:15:22.522451448Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: failed to load CNI config list file /etc/cni/net.d/10-calico.conflist: error parsing configuration list: unexpected end of JSON input: invalid cni config: failed to load cni config" May 15 15:15:22.543913 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3b54bbed567c9f640667f55f5f36dd4d9107d0ec2bb414e75cbcee8bab3e2584-rootfs.mount: Deactivated successfully. May 15 15:15:23.325074 kubelet[2694]: E0515 15:15:23.324647 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:23.352963 containerd[1549]: time="2025-05-15T15:15:23.351449821Z" level=info msg="CreateContainer within sandbox \"0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 15 15:15:23.372084 systemd[1]: Started sshd@28-24.144.90.91:22-139.178.68.195:54158.service - OpenSSH per-connection server daemon (139.178.68.195:54158). May 15 15:15:23.377030 containerd[1549]: time="2025-05-15T15:15:23.376018710Z" level=info msg="Container 7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e: CDI devices from CRI Config.CDIDevices: []" May 15 15:15:23.386781 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2408416415.mount: Deactivated successfully. May 15 15:15:23.395103 containerd[1549]: time="2025-05-15T15:15:23.394793709Z" level=info msg="CreateContainer within sandbox \"0ecfd8078a5416cce13c9aff743e74dc6afed462ca3228c62c0a2902ff4c73be\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e\"" May 15 15:15:23.398772 containerd[1549]: time="2025-05-15T15:15:23.397648043Z" level=info msg="StartContainer for \"7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e\"" May 15 15:15:23.400820 containerd[1549]: time="2025-05-15T15:15:23.400777995Z" level=info msg="connecting to shim 7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e" address="unix:///run/containerd/s/7456a8ec0c8096ff76a14ff380fc2b6f2154c56fcfb429960d91b5c911aa7730" protocol=ttrpc version=3 May 15 15:15:23.430383 systemd[1]: Started cri-containerd-7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e.scope - libcontainer container 7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e. May 15 15:15:23.535014 sshd[7020]: Accepted publickey for core from 139.178.68.195 port 54158 ssh2: RSA SHA256:MR6P4SMnBj7Bljnyb1daa15ne/ebNhdFSQPikHCJ1Fk May 15 15:15:23.539747 sshd-session[7020]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:15:23.541107 containerd[1549]: time="2025-05-15T15:15:23.540255420Z" level=info msg="StartContainer for \"7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e\" returns successfully" May 15 15:15:23.561373 systemd-logind[1527]: New session 26 of user core. May 15 15:15:23.570262 systemd[1]: Started session-26.scope - Session 26 of User core. May 15 15:15:24.274847 sshd[7056]: Connection closed by 139.178.68.195 port 54158 May 15 15:15:24.275890 sshd-session[7020]: pam_unix(sshd:session): session closed for user core May 15 15:15:24.280882 systemd[1]: sshd@28-24.144.90.91:22-139.178.68.195:54158.service: Deactivated successfully. May 15 15:15:24.284804 systemd[1]: session-26.scope: Deactivated successfully. May 15 15:15:24.287001 systemd-logind[1527]: Session 26 logged out. Waiting for processes to exit. May 15 15:15:24.288662 systemd-logind[1527]: Removed session 26. May 15 15:15:24.333794 kubelet[2694]: E0515 15:15:24.333751 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:24.377511 kubelet[2694]: I0515 15:15:24.376755 2694 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-ls2dl" podStartSLOduration=4.376699051 podStartE2EDuration="4.376699051s" podCreationTimestamp="2025-05-15 15:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 15:15:24.37525334 +0000 UTC m=+118.056126388" watchObservedRunningTime="2025-05-15 15:15:24.376699051 +0000 UTC m=+118.057572079" May 15 15:15:24.448964 containerd[1549]: time="2025-05-15T15:15:24.448878417Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e\" id:\"2d21018158f50eae59ed0b18e8f6d25a827335746de07e96bb1794671a14a410\" pid:7101 exit_status:1 exited_at:{seconds:1747322124 nanos:448269627}" May 15 15:15:25.339028 kubelet[2694]: E0515 15:15:25.338659 2694 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" May 15 15:15:25.504308 containerd[1549]: time="2025-05-15T15:15:25.504257179Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7c4f2c31f57a7b17c8ade304239797862c45cefed37d0b38eb6bf241e853792e\" id:\"6ee54e966c75a7650acb302aa1b1762abae9c0b51fc75e7feaa02bba8dd7a99e\" pid:7241 exit_status:1 exited_at:{seconds:1747322125 nanos:503641048}" May 15 15:15:26.953484 containerd[1549]: time="2025-05-15T15:15:26.953396070Z" level=info msg="StopPodSandbox for \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\"" May 15 15:15:26.954790 containerd[1549]: time="2025-05-15T15:15:26.954408470Z" level=info msg="TearDown network for sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" successfully" May 15 15:15:26.954790 containerd[1549]: time="2025-05-15T15:15:26.954446589Z" level=info msg="StopPodSandbox for \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" returns successfully" May 15 15:15:26.955595 containerd[1549]: time="2025-05-15T15:15:26.954971500Z" level=info msg="RemovePodSandbox for \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\"" May 15 15:15:26.955595 containerd[1549]: time="2025-05-15T15:15:26.954999265Z" level=info msg="Forcibly stopping sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\"" May 15 15:15:26.955595 containerd[1549]: time="2025-05-15T15:15:26.955268030Z" level=info msg="TearDown network for sandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" successfully" May 15 15:15:26.957494 containerd[1549]: time="2025-05-15T15:15:26.957453049Z" level=info msg="Ensure that sandbox 887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82 in task-service has been cleanup successfully" May 15 15:15:26.960110 containerd[1549]: time="2025-05-15T15:15:26.960027073Z" level=info msg="RemovePodSandbox \"887ad25f6ed90bed250f40b607eead3ef2e41c3ce27477d9faa078a62bffec82\" returns successfully" May 15 15:15:26.961028 containerd[1549]: time="2025-05-15T15:15:26.960924648Z" level=info msg="StopPodSandbox for \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\"" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.021 [WARNING][7344] cni-plugin/k8s.go 566: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.022 [INFO][7344] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.022 [INFO][7344] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" iface="eth0" netns="" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.022 [INFO][7344] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.022 [INFO][7344] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.054 [INFO][7352] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.054 [INFO][7352] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.054 [INFO][7352] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.062 [WARNING][7352] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.063 [INFO][7352] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.065 [INFO][7352] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:15:27.071830 containerd[1549]: 2025-05-15 15:15:27.068 [INFO][7344] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:15:27.072735 containerd[1549]: time="2025-05-15T15:15:27.072442128Z" level=info msg="TearDown network for sandbox \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" successfully" May 15 15:15:27.072735 containerd[1549]: time="2025-05-15T15:15:27.072487287Z" level=info msg="StopPodSandbox for \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" returns successfully" May 15 15:15:27.073681 containerd[1549]: time="2025-05-15T15:15:27.073275526Z" level=info msg="RemovePodSandbox for \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\"" May 15 15:15:27.073681 containerd[1549]: time="2025-05-15T15:15:27.073307195Z" level=info msg="Forcibly stopping sandbox \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\"" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.119 [WARNING][7371] cni-plugin/k8s.go 566: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.119 [INFO][7371] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.119 [INFO][7371] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" iface="eth0" netns="" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.119 [INFO][7371] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.119 [INFO][7371] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.150 [INFO][7378] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.150 [INFO][7378] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.150 [INFO][7378] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.159 [WARNING][7378] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.159 [INFO][7378] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" HandleID="k8s-pod-network.41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--apiserver--585b9b4df4--cr2q4-eth0" May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.161 [INFO][7378] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 15:15:27.166536 containerd[1549]: 2025-05-15 15:15:27.164 [INFO][7371] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611" May 15 15:15:27.167291 containerd[1549]: time="2025-05-15T15:15:27.167164181Z" level=info msg="TearDown network for sandbox \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" successfully" May 15 15:15:27.173478 containerd[1549]: time="2025-05-15T15:15:27.173210042Z" level=info msg="Ensure that sandbox 41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611 in task-service has been cleanup successfully" May 15 15:15:27.176329 containerd[1549]: time="2025-05-15T15:15:27.176290621Z" level=info msg="RemovePodSandbox \"41860d3ed5e5e73652718163550ca5755af9a2c8b04b3030a258885adfada611\" returns successfully" May 15 15:15:27.177103 containerd[1549]: time="2025-05-15T15:15:27.177062443Z" level=info msg="StopPodSandbox for \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\"" May 15 15:15:27.177289 containerd[1549]: time="2025-05-15T15:15:27.177218007Z" level=info msg="TearDown network for sandbox \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" successfully" May 15 15:15:27.177289 containerd[1549]: time="2025-05-15T15:15:27.177248736Z" level=info msg="StopPodSandbox for \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" returns successfully" May 15 15:15:27.178299 containerd[1549]: time="2025-05-15T15:15:27.178069848Z" level=info msg="RemovePodSandbox for \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\"" May 15 15:15:27.178299 containerd[1549]: time="2025-05-15T15:15:27.178105945Z" level=info msg="Forcibly stopping sandbox \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\"" May 15 15:15:27.178299 containerd[1549]: time="2025-05-15T15:15:27.178215721Z" level=info msg="TearDown network for sandbox \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" successfully" May 15 15:15:27.180604 containerd[1549]: time="2025-05-15T15:15:27.180563795Z" level=info msg="Ensure that sandbox 4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58 in task-service has been cleanup successfully" May 15 15:15:27.185112 containerd[1549]: time="2025-05-15T15:15:27.184983020Z" level=info msg="RemovePodSandbox \"4265131a49e8a8f988626a1973e8861de1f1c5e430592135fcf53432de1b7c58\" returns successfully" May 15 15:15:27.186123 containerd[1549]: time="2025-05-15T15:15:27.185634075Z" level=info msg="StopPodSandbox for \"a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f\"" May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.248 [WARNING][7397] cni-plugin/k8s.go 566: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" WorkloadEndpoint="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.249 [INFO][7397] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.249 [INFO][7397] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" iface="eth0" netns="" May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.249 [INFO][7397] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.249 [INFO][7397] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.277 [INFO][7404] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" HandleID="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0" May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.277 [INFO][7404] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.277 [INFO][7404] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 15:15:27.295135 containerd[1549]: 2025-05-15 15:15:27.286 [WARNING][7404] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" HandleID="k8s-pod-network.a531f941ba13d3e3496c6ac4497768100dd2090ea02c1744c6c4e778630a7f7f" Workload="ci--4334.0.0--a--073c24674a-k8s-calico--kube--controllers--c8cd4b68c--85hmc-eth0"