Apr 24 23:33:16.960098 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Apr 24 22:11:38 -00 2026 Apr 24 23:33:16.960115 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=c8442747465ed99a522e07b8746f6a7817fb39c2025d7438698e3b90e9c0defb Apr 24 23:33:16.960125 kernel: BIOS-provided physical RAM map: Apr 24 23:33:16.960130 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Apr 24 23:33:16.960134 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ed3efff] usable Apr 24 23:33:16.960138 kernel: BIOS-e820: [mem 0x000000007ed3f000-0x000000007edfffff] reserved Apr 24 23:33:16.960143 kernel: BIOS-e820: [mem 0x000000007ee00000-0x000000007f8ecfff] usable Apr 24 23:33:16.960148 kernel: BIOS-e820: [mem 0x000000007f8ed000-0x000000007f9ecfff] reserved Apr 24 23:33:16.960152 kernel: BIOS-e820: [mem 0x000000007f9ed000-0x000000007faecfff] type 20 Apr 24 23:33:16.960157 kernel: BIOS-e820: [mem 0x000000007faed000-0x000000007fb6cfff] reserved Apr 24 23:33:16.960162 kernel: BIOS-e820: [mem 0x000000007fb6d000-0x000000007fb7efff] ACPI data Apr 24 23:33:16.960169 kernel: BIOS-e820: [mem 0x000000007fb7f000-0x000000007fbfefff] ACPI NVS Apr 24 23:33:16.960173 kernel: BIOS-e820: [mem 0x000000007fbff000-0x000000007ff7bfff] usable Apr 24 23:33:16.960177 kernel: BIOS-e820: [mem 0x000000007ff7c000-0x000000007fffffff] reserved Apr 24 23:33:16.960183 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Apr 24 23:33:16.960187 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Apr 24 23:33:16.960194 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Apr 24 23:33:16.960199 kernel: BIOS-e820: [mem 0x0000000100000000-0x0000000179ffffff] usable Apr 24 23:33:16.960203 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Apr 24 23:33:16.960208 kernel: NX (Execute Disable) protection: active Apr 24 23:33:16.960213 kernel: APIC: Static calls initialized Apr 24 23:33:16.960217 kernel: efi: EFI v2.7 by Ubuntu distribution of EDK II Apr 24 23:33:16.960222 kernel: efi: SMBIOS=0x7f988000 SMBIOS 3.0=0x7f986000 ACPI=0x7fb7e000 ACPI 2.0=0x7fb7e014 MEMATTR=0x7e845198 Apr 24 23:33:16.960226 kernel: efi: Remove mem135: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Apr 24 23:33:16.960231 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Apr 24 23:33:16.960236 kernel: SMBIOS 3.0.0 present. Apr 24 23:33:16.960240 kernel: DMI: Hetzner vServer/Standard PC (Q35 + ICH9, 2009), BIOS 20171111 11/11/2017 Apr 24 23:33:16.960245 kernel: Hypervisor detected: KVM Apr 24 23:33:16.960252 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Apr 24 23:33:16.960257 kernel: kvm-clock: using sched offset of 12789464182 cycles Apr 24 23:33:16.960261 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Apr 24 23:33:16.960266 kernel: tsc: Detected 2399.998 MHz processor Apr 24 23:33:16.960271 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Apr 24 23:33:16.960276 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Apr 24 23:33:16.960281 kernel: last_pfn = 0x17a000 max_arch_pfn = 0x10000000000 Apr 24 23:33:16.960285 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Apr 24 23:33:16.960290 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Apr 24 23:33:16.960297 kernel: last_pfn = 0x7ff7c max_arch_pfn = 0x10000000000 Apr 24 23:33:16.960302 kernel: Using GB pages for direct mapping Apr 24 23:33:16.960306 kernel: Secure boot disabled Apr 24 23:33:16.960314 kernel: ACPI: Early table checksum verification disabled Apr 24 23:33:16.960319 kernel: ACPI: RSDP 0x000000007FB7E014 000024 (v02 BOCHS ) Apr 24 23:33:16.960324 kernel: ACPI: XSDT 0x000000007FB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Apr 24 23:33:16.960329 kernel: ACPI: FACP 0x000000007FB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 24 23:33:16.960336 kernel: ACPI: DSDT 0x000000007FB7A000 002443 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 24 23:33:16.960341 kernel: ACPI: FACS 0x000000007FBDD000 000040 Apr 24 23:33:16.960346 kernel: ACPI: APIC 0x000000007FB78000 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 24 23:33:16.960351 kernel: ACPI: HPET 0x000000007FB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 24 23:33:16.960356 kernel: ACPI: MCFG 0x000000007FB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 24 23:33:16.960360 kernel: ACPI: WAET 0x000000007FB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 24 23:33:16.960365 kernel: ACPI: BGRT 0x000000007FB74000 000038 (v01 INTEL EDK2 00000002 01000013) Apr 24 23:33:16.960373 kernel: ACPI: Reserving FACP table memory at [mem 0x7fb79000-0x7fb790f3] Apr 24 23:33:16.960378 kernel: ACPI: Reserving DSDT table memory at [mem 0x7fb7a000-0x7fb7c442] Apr 24 23:33:16.960383 kernel: ACPI: Reserving FACS table memory at [mem 0x7fbdd000-0x7fbdd03f] Apr 24 23:33:16.960387 kernel: ACPI: Reserving APIC table memory at [mem 0x7fb78000-0x7fb7807f] Apr 24 23:33:16.960392 kernel: ACPI: Reserving HPET table memory at [mem 0x7fb77000-0x7fb77037] Apr 24 23:33:16.960397 kernel: ACPI: Reserving MCFG table memory at [mem 0x7fb76000-0x7fb7603b] Apr 24 23:33:16.960402 kernel: ACPI: Reserving WAET table memory at [mem 0x7fb75000-0x7fb75027] Apr 24 23:33:16.960407 kernel: ACPI: Reserving BGRT table memory at [mem 0x7fb74000-0x7fb74037] Apr 24 23:33:16.960412 kernel: No NUMA configuration found Apr 24 23:33:16.960419 kernel: Faking a node at [mem 0x0000000000000000-0x0000000179ffffff] Apr 24 23:33:16.960424 kernel: NODE_DATA(0) allocated [mem 0x179ff8000-0x179ffdfff] Apr 24 23:33:16.960429 kernel: Zone ranges: Apr 24 23:33:16.960434 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Apr 24 23:33:16.960439 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Apr 24 23:33:16.960444 kernel: Normal [mem 0x0000000100000000-0x0000000179ffffff] Apr 24 23:33:16.960448 kernel: Movable zone start for each node Apr 24 23:33:16.960453 kernel: Early memory node ranges Apr 24 23:33:16.960458 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Apr 24 23:33:16.960463 kernel: node 0: [mem 0x0000000000100000-0x000000007ed3efff] Apr 24 23:33:16.960471 kernel: node 0: [mem 0x000000007ee00000-0x000000007f8ecfff] Apr 24 23:33:16.960475 kernel: node 0: [mem 0x000000007fbff000-0x000000007ff7bfff] Apr 24 23:33:16.960480 kernel: node 0: [mem 0x0000000100000000-0x0000000179ffffff] Apr 24 23:33:16.960485 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x0000000179ffffff] Apr 24 23:33:16.960490 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Apr 24 23:33:16.960495 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Apr 24 23:33:16.960500 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Apr 24 23:33:16.960504 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Apr 24 23:33:16.960509 kernel: On node 0, zone Normal: 132 pages in unavailable ranges Apr 24 23:33:16.960517 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Apr 24 23:33:16.960522 kernel: ACPI: PM-Timer IO Port: 0x608 Apr 24 23:33:16.960527 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Apr 24 23:33:16.960532 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Apr 24 23:33:16.960537 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Apr 24 23:33:16.960541 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Apr 24 23:33:16.960546 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Apr 24 23:33:16.960551 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Apr 24 23:33:16.960556 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Apr 24 23:33:16.960564 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Apr 24 23:33:16.960568 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Apr 24 23:33:16.960573 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Apr 24 23:33:16.960578 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Apr 24 23:33:16.960583 kernel: [mem 0x80000000-0xdfffffff] available for PCI devices Apr 24 23:33:16.960588 kernel: Booting paravirtualized kernel on KVM Apr 24 23:33:16.960593 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Apr 24 23:33:16.960598 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Apr 24 23:33:16.960603 kernel: percpu: Embedded 57 pages/cpu s196328 r8192 d28952 u1048576 Apr 24 23:33:16.960610 kernel: pcpu-alloc: s196328 r8192 d28952 u1048576 alloc=1*2097152 Apr 24 23:33:16.960615 kernel: pcpu-alloc: [0] 0 1 Apr 24 23:33:16.960620 kernel: kvm-guest: PV spinlocks disabled, no host support Apr 24 23:33:16.960626 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=c8442747465ed99a522e07b8746f6a7817fb39c2025d7438698e3b90e9c0defb Apr 24 23:33:16.960631 kernel: random: crng init done Apr 24 23:33:16.960636 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 24 23:33:16.960641 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 24 23:33:16.960646 kernel: Fallback order for Node 0: 0 Apr 24 23:33:16.960653 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1004632 Apr 24 23:33:16.960658 kernel: Policy zone: Normal Apr 24 23:33:16.960663 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 24 23:33:16.960668 kernel: software IO TLB: area num 2. Apr 24 23:33:16.960673 kernel: Memory: 3827828K/4091168K available (12288K kernel code, 2288K rwdata, 22752K rodata, 42896K init, 2300K bss, 263136K reserved, 0K cma-reserved) Apr 24 23:33:16.960678 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 24 23:33:16.960683 kernel: ftrace: allocating 37996 entries in 149 pages Apr 24 23:33:16.960687 kernel: ftrace: allocated 149 pages with 4 groups Apr 24 23:33:16.960692 kernel: Dynamic Preempt: voluntary Apr 24 23:33:16.960700 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 24 23:33:16.960708 kernel: rcu: RCU event tracing is enabled. Apr 24 23:33:16.960713 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 24 23:33:16.960718 kernel: Trampoline variant of Tasks RCU enabled. Apr 24 23:33:16.960730 kernel: Rude variant of Tasks RCU enabled. Apr 24 23:33:16.960737 kernel: Tracing variant of Tasks RCU enabled. Apr 24 23:33:16.960743 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 24 23:33:16.960748 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 24 23:33:16.960753 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Apr 24 23:33:16.960758 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 24 23:33:16.960763 kernel: Console: colour dummy device 80x25 Apr 24 23:33:16.960768 kernel: printk: console [tty0] enabled Apr 24 23:33:16.960776 kernel: printk: console [ttyS0] enabled Apr 24 23:33:16.960781 kernel: ACPI: Core revision 20230628 Apr 24 23:33:16.960786 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Apr 24 23:33:16.960791 kernel: APIC: Switch to symmetric I/O mode setup Apr 24 23:33:16.960796 kernel: x2apic enabled Apr 24 23:33:16.960804 kernel: APIC: Switched APIC routing to: physical x2apic Apr 24 23:33:16.960809 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Apr 24 23:33:16.960814 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Apr 24 23:33:16.960819 kernel: Calibrating delay loop (skipped) preset value.. 4799.99 BogoMIPS (lpj=2399998) Apr 24 23:33:16.960825 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Apr 24 23:33:16.960830 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Apr 24 23:33:16.960835 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Apr 24 23:33:16.960840 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Apr 24 23:33:16.960845 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Apr 24 23:33:16.960852 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Apr 24 23:33:16.960857 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Apr 24 23:33:16.960863 kernel: active return thunk: srso_alias_return_thunk Apr 24 23:33:16.960868 kernel: Speculative Return Stack Overflow: Mitigation: Safe RET Apr 24 23:33:16.960873 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Apr 24 23:33:16.960878 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Apr 24 23:33:16.960883 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Apr 24 23:33:16.960888 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Apr 24 23:33:16.960893 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Apr 24 23:33:16.960901 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Apr 24 23:33:16.960906 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Apr 24 23:33:16.960911 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Apr 24 23:33:16.960916 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Apr 24 23:33:16.960921 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Apr 24 23:33:16.960927 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Apr 24 23:33:16.960932 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Apr 24 23:33:16.960937 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Apr 24 23:33:16.960942 kernel: x86/fpu: xstate_offset[9]: 2432, xstate_sizes[9]: 8 Apr 24 23:33:16.960977 kernel: x86/fpu: Enabled xstate features 0x2e7, context size is 2440 bytes, using 'compacted' format. Apr 24 23:33:16.960984 kernel: Freeing SMP alternatives memory: 32K Apr 24 23:33:16.960989 kernel: pid_max: default: 32768 minimum: 301 Apr 24 23:33:16.960994 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 24 23:33:16.960999 kernel: landlock: Up and running. Apr 24 23:33:16.961004 kernel: SELinux: Initializing. Apr 24 23:33:16.961010 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 24 23:33:16.961015 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 24 23:33:16.961020 kernel: smpboot: CPU0: AMD EPYC-Genoa Processor (family: 0x19, model: 0x11, stepping: 0x0) Apr 24 23:33:16.961028 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 24 23:33:16.961033 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 24 23:33:16.961038 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 24 23:33:16.961043 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Apr 24 23:33:16.961048 kernel: ... version: 0 Apr 24 23:33:16.961053 kernel: ... bit width: 48 Apr 24 23:33:16.961058 kernel: ... generic registers: 6 Apr 24 23:33:16.961063 kernel: ... value mask: 0000ffffffffffff Apr 24 23:33:16.961071 kernel: ... max period: 00007fffffffffff Apr 24 23:33:16.961076 kernel: ... fixed-purpose events: 0 Apr 24 23:33:16.961081 kernel: ... event mask: 000000000000003f Apr 24 23:33:16.961086 kernel: signal: max sigframe size: 3376 Apr 24 23:33:16.961091 kernel: rcu: Hierarchical SRCU implementation. Apr 24 23:33:16.961097 kernel: rcu: Max phase no-delay instances is 400. Apr 24 23:33:16.961102 kernel: smp: Bringing up secondary CPUs ... Apr 24 23:33:16.961107 kernel: smpboot: x86: Booting SMP configuration: Apr 24 23:33:16.961112 kernel: .... node #0, CPUs: #1 Apr 24 23:33:16.961117 kernel: smp: Brought up 1 node, 2 CPUs Apr 24 23:33:16.961124 kernel: smpboot: Max logical packages: 1 Apr 24 23:33:16.961130 kernel: smpboot: Total of 2 processors activated (9599.99 BogoMIPS) Apr 24 23:33:16.961135 kernel: devtmpfs: initialized Apr 24 23:33:16.961140 kernel: x86/mm: Memory block size: 128MB Apr 24 23:33:16.961145 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7fb7f000-0x7fbfefff] (524288 bytes) Apr 24 23:33:16.961150 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 24 23:33:16.961155 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 24 23:33:16.961160 kernel: pinctrl core: initialized pinctrl subsystem Apr 24 23:33:16.961166 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 24 23:33:16.961173 kernel: audit: initializing netlink subsys (disabled) Apr 24 23:33:16.961178 kernel: audit: type=2000 audit(1777073596.009:1): state=initialized audit_enabled=0 res=1 Apr 24 23:33:16.961183 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 24 23:33:16.961188 kernel: thermal_sys: Registered thermal governor 'user_space' Apr 24 23:33:16.961193 kernel: cpuidle: using governor menu Apr 24 23:33:16.961199 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 24 23:33:16.961204 kernel: dca service started, version 1.12.1 Apr 24 23:33:16.961209 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Apr 24 23:33:16.961214 kernel: PCI: Using configuration type 1 for base access Apr 24 23:33:16.961221 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Apr 24 23:33:16.961227 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 24 23:33:16.961232 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Apr 24 23:33:16.961237 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 24 23:33:16.961242 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Apr 24 23:33:16.961247 kernel: ACPI: Added _OSI(Module Device) Apr 24 23:33:16.961253 kernel: ACPI: Added _OSI(Processor Device) Apr 24 23:33:16.961258 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 24 23:33:16.961263 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 24 23:33:16.961270 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Apr 24 23:33:16.961275 kernel: ACPI: Interpreter enabled Apr 24 23:33:16.961280 kernel: ACPI: PM: (supports S0 S5) Apr 24 23:33:16.961285 kernel: ACPI: Using IOAPIC for interrupt routing Apr 24 23:33:16.961291 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Apr 24 23:33:16.961296 kernel: PCI: Using E820 reservations for host bridge windows Apr 24 23:33:16.961301 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Apr 24 23:33:16.961306 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 24 23:33:16.961469 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 24 23:33:16.961580 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Apr 24 23:33:16.961680 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Apr 24 23:33:16.961688 kernel: PCI host bridge to bus 0000:00 Apr 24 23:33:16.961791 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Apr 24 23:33:16.961898 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Apr 24 23:33:16.962015 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Apr 24 23:33:16.962126 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xdfffffff window] Apr 24 23:33:16.962218 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Apr 24 23:33:16.962305 kernel: pci_bus 0000:00: root bus resource [mem 0xc000000000-0xc7ffffffff window] Apr 24 23:33:16.962391 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 24 23:33:16.962501 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Apr 24 23:33:16.962607 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x030000 Apr 24 23:33:16.962708 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80000000-0x807fffff pref] Apr 24 23:33:16.962805 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc060500000-0xc060503fff 64bit pref] Apr 24 23:33:16.962901 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8138a000-0x8138afff] Apr 24 23:33:16.965045 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Apr 24 23:33:16.965155 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Apr 24 23:33:16.965256 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Apr 24 23:33:16.965362 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.965463 kernel: pci 0000:00:02.0: reg 0x10: [mem 0x81389000-0x81389fff] Apr 24 23:33:16.965568 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.965665 kernel: pci 0000:00:02.1: reg 0x10: [mem 0x81388000-0x81388fff] Apr 24 23:33:16.965767 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.965862 kernel: pci 0000:00:02.2: reg 0x10: [mem 0x81387000-0x81387fff] Apr 24 23:33:16.965991 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.966095 kernel: pci 0000:00:02.3: reg 0x10: [mem 0x81386000-0x81386fff] Apr 24 23:33:16.966197 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.966292 kernel: pci 0000:00:02.4: reg 0x10: [mem 0x81385000-0x81385fff] Apr 24 23:33:16.966393 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.966489 kernel: pci 0000:00:02.5: reg 0x10: [mem 0x81384000-0x81384fff] Apr 24 23:33:16.966591 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.966689 kernel: pci 0000:00:02.6: reg 0x10: [mem 0x81383000-0x81383fff] Apr 24 23:33:16.966791 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.966887 kernel: pci 0000:00:02.7: reg 0x10: [mem 0x81382000-0x81382fff] Apr 24 23:33:16.969059 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Apr 24 23:33:16.969202 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x81381000-0x81381fff] Apr 24 23:33:16.969354 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Apr 24 23:33:16.969492 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Apr 24 23:33:16.969639 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Apr 24 23:33:16.969783 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x6040-0x605f] Apr 24 23:33:16.969917 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0x81380000-0x81380fff] Apr 24 23:33:16.971124 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Apr 24 23:33:16.971272 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6000-0x603f] Apr 24 23:33:16.971440 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Apr 24 23:33:16.972091 kernel: pci 0000:01:00.0: reg 0x14: [mem 0x81200000-0x81200fff] Apr 24 23:33:16.972235 kernel: pci 0000:01:00.0: reg 0x20: [mem 0xc060000000-0xc060003fff 64bit pref] Apr 24 23:33:16.972372 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Apr 24 23:33:16.972510 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Apr 24 23:33:16.972647 kernel: pci 0000:00:02.0: bridge window [mem 0x81200000-0x812fffff] Apr 24 23:33:16.972775 kernel: pci 0000:00:02.0: bridge window [mem 0xc060000000-0xc0600fffff 64bit pref] Apr 24 23:33:16.972928 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Apr 24 23:33:16.973103 kernel: pci 0000:02:00.0: reg 0x10: [mem 0x81100000-0x81103fff 64bit] Apr 24 23:33:16.973250 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Apr 24 23:33:16.973392 kernel: pci 0000:00:02.1: bridge window [mem 0x81100000-0x811fffff] Apr 24 23:33:16.973550 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Apr 24 23:33:16.973702 kernel: pci 0000:03:00.0: reg 0x14: [mem 0x81000000-0x81000fff] Apr 24 23:33:16.973843 kernel: pci 0000:03:00.0: reg 0x20: [mem 0xc060100000-0xc060103fff 64bit pref] Apr 24 23:33:16.976061 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Apr 24 23:33:16.976215 kernel: pci 0000:00:02.2: bridge window [mem 0x81000000-0x810fffff] Apr 24 23:33:16.976358 kernel: pci 0000:00:02.2: bridge window [mem 0xc060100000-0xc0601fffff 64bit pref] Apr 24 23:33:16.976511 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Apr 24 23:33:16.976654 kernel: pci 0000:04:00.0: reg 0x20: [mem 0xc060200000-0xc060203fff 64bit pref] Apr 24 23:33:16.976789 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Apr 24 23:33:16.976925 kernel: pci 0000:00:02.3: bridge window [mem 0xc060200000-0xc0602fffff 64bit pref] Apr 24 23:33:16.977090 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Apr 24 23:33:16.979292 kernel: pci 0000:05:00.0: reg 0x14: [mem 0x80f00000-0x80f00fff] Apr 24 23:33:16.979429 kernel: pci 0000:05:00.0: reg 0x20: [mem 0xc060300000-0xc060303fff 64bit pref] Apr 24 23:33:16.979558 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Apr 24 23:33:16.979669 kernel: pci 0000:00:02.4: bridge window [mem 0x80f00000-0x80ffffff] Apr 24 23:33:16.979766 kernel: pci 0000:00:02.4: bridge window [mem 0xc060300000-0xc0603fffff 64bit pref] Apr 24 23:33:16.979876 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Apr 24 23:33:16.982226 kernel: pci 0000:06:00.0: reg 0x14: [mem 0x80e00000-0x80e00fff] Apr 24 23:33:16.982390 kernel: pci 0000:06:00.0: reg 0x20: [mem 0xc060400000-0xc060403fff 64bit pref] Apr 24 23:33:16.982526 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Apr 24 23:33:16.982662 kernel: pci 0000:00:02.5: bridge window [mem 0x80e00000-0x80efffff] Apr 24 23:33:16.982789 kernel: pci 0000:00:02.5: bridge window [mem 0xc060400000-0xc0604fffff 64bit pref] Apr 24 23:33:16.982801 kernel: acpiphp: Slot [0] registered Apr 24 23:33:16.982962 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Apr 24 23:33:16.983102 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x80c00000-0x80c00fff] Apr 24 23:33:16.983239 kernel: pci 0000:07:00.0: reg 0x20: [mem 0xc000000000-0xc000003fff 64bit pref] Apr 24 23:33:16.983365 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Apr 24 23:33:16.983483 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Apr 24 23:33:16.983605 kernel: pci 0000:00:02.6: bridge window [mem 0x80c00000-0x80dfffff] Apr 24 23:33:16.983733 kernel: pci 0000:00:02.6: bridge window [mem 0xc000000000-0xc01fffffff 64bit pref] Apr 24 23:33:16.983741 kernel: acpiphp: Slot [0-2] registered Apr 24 23:33:16.983873 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Apr 24 23:33:16.984052 kernel: pci 0000:00:02.7: bridge window [mem 0x80a00000-0x80bfffff] Apr 24 23:33:16.984194 kernel: pci 0000:00:02.7: bridge window [mem 0xc020000000-0xc03fffffff 64bit pref] Apr 24 23:33:16.984206 kernel: acpiphp: Slot [0-3] registered Apr 24 23:33:16.984342 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Apr 24 23:33:16.984476 kernel: pci 0000:00:03.0: bridge window [mem 0x80800000-0x809fffff] Apr 24 23:33:16.984608 kernel: pci 0000:00:03.0: bridge window [mem 0xc040000000-0xc05fffffff 64bit pref] Apr 24 23:33:16.984619 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Apr 24 23:33:16.984627 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Apr 24 23:33:16.984635 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Apr 24 23:33:16.984643 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Apr 24 23:33:16.984656 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Apr 24 23:33:16.984662 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Apr 24 23:33:16.984668 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Apr 24 23:33:16.984674 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Apr 24 23:33:16.984683 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Apr 24 23:33:16.984692 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Apr 24 23:33:16.984700 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Apr 24 23:33:16.984707 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Apr 24 23:33:16.984714 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Apr 24 23:33:16.984726 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Apr 24 23:33:16.984735 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Apr 24 23:33:16.984742 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Apr 24 23:33:16.984751 kernel: iommu: Default domain type: Translated Apr 24 23:33:16.984760 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Apr 24 23:33:16.984768 kernel: efivars: Registered efivars operations Apr 24 23:33:16.984776 kernel: PCI: Using ACPI for IRQ routing Apr 24 23:33:16.984785 kernel: PCI: pci_cache_line_size set to 64 bytes Apr 24 23:33:16.984794 kernel: e820: reserve RAM buffer [mem 0x7ed3f000-0x7fffffff] Apr 24 23:33:16.984804 kernel: e820: reserve RAM buffer [mem 0x7f8ed000-0x7fffffff] Apr 24 23:33:16.984813 kernel: e820: reserve RAM buffer [mem 0x7ff7c000-0x7fffffff] Apr 24 23:33:16.984821 kernel: e820: reserve RAM buffer [mem 0x17a000000-0x17bffffff] Apr 24 23:33:16.986027 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Apr 24 23:33:16.986175 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Apr 24 23:33:16.986312 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Apr 24 23:33:16.986323 kernel: vgaarb: loaded Apr 24 23:33:16.986332 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Apr 24 23:33:16.986341 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Apr 24 23:33:16.986355 kernel: clocksource: Switched to clocksource kvm-clock Apr 24 23:33:16.986364 kernel: VFS: Disk quotas dquot_6.6.0 Apr 24 23:33:16.986373 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 24 23:33:16.986382 kernel: pnp: PnP ACPI init Apr 24 23:33:16.986533 kernel: system 00:04: [mem 0xe0000000-0xefffffff window] has been reserved Apr 24 23:33:16.986546 kernel: pnp: PnP ACPI: found 5 devices Apr 24 23:33:16.986555 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Apr 24 23:33:16.986564 kernel: NET: Registered PF_INET protocol family Apr 24 23:33:16.986594 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 24 23:33:16.986605 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 24 23:33:16.986615 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 24 23:33:16.986623 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 24 23:33:16.986632 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 24 23:33:16.986641 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 24 23:33:16.986650 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 24 23:33:16.986658 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 24 23:33:16.986670 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 24 23:33:16.986679 kernel: NET: Registered PF_XDP protocol family Apr 24 23:33:16.986832 kernel: pci 0000:01:00.0: can't claim BAR 6 [mem 0xfff80000-0xffffffff pref]: no compatible bridge window Apr 24 23:33:16.987728 kernel: pci 0000:07:00.0: can't claim BAR 6 [mem 0xfff80000-0xffffffff pref]: no compatible bridge window Apr 24 23:33:16.987880 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Apr 24 23:33:16.988054 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Apr 24 23:33:16.988198 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Apr 24 23:33:16.988347 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x1000-0x1fff] Apr 24 23:33:16.988484 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x2000-0x2fff] Apr 24 23:33:16.988629 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x3000-0x3fff] Apr 24 23:33:16.988775 kernel: pci 0000:01:00.0: BAR 6: assigned [mem 0x81280000-0x812fffff pref] Apr 24 23:33:16.988895 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Apr 24 23:33:16.989043 kernel: pci 0000:00:02.0: bridge window [mem 0x81200000-0x812fffff] Apr 24 23:33:16.989143 kernel: pci 0000:00:02.0: bridge window [mem 0xc060000000-0xc0600fffff 64bit pref] Apr 24 23:33:16.989257 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Apr 24 23:33:16.989381 kernel: pci 0000:00:02.1: bridge window [mem 0x81100000-0x811fffff] Apr 24 23:33:16.989489 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Apr 24 23:33:16.989585 kernel: pci 0000:00:02.2: bridge window [mem 0x81000000-0x810fffff] Apr 24 23:33:16.989683 kernel: pci 0000:00:02.2: bridge window [mem 0xc060100000-0xc0601fffff 64bit pref] Apr 24 23:33:16.989781 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Apr 24 23:33:16.989882 kernel: pci 0000:00:02.3: bridge window [mem 0xc060200000-0xc0602fffff 64bit pref] Apr 24 23:33:16.994082 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Apr 24 23:33:16.994194 kernel: pci 0000:00:02.4: bridge window [mem 0x80f00000-0x80ffffff] Apr 24 23:33:16.994294 kernel: pci 0000:00:02.4: bridge window [mem 0xc060300000-0xc0603fffff 64bit pref] Apr 24 23:33:16.994393 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Apr 24 23:33:16.994489 kernel: pci 0000:00:02.5: bridge window [mem 0x80e00000-0x80efffff] Apr 24 23:33:16.994585 kernel: pci 0000:00:02.5: bridge window [mem 0xc060400000-0xc0604fffff 64bit pref] Apr 24 23:33:16.994686 kernel: pci 0000:07:00.0: BAR 6: assigned [mem 0x80c80000-0x80cfffff pref] Apr 24 23:33:16.994787 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Apr 24 23:33:16.994882 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff] Apr 24 23:33:16.994996 kernel: pci 0000:00:02.6: bridge window [mem 0x80c00000-0x80dfffff] Apr 24 23:33:16.995093 kernel: pci 0000:00:02.6: bridge window [mem 0xc000000000-0xc01fffffff 64bit pref] Apr 24 23:33:16.995188 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Apr 24 23:33:16.995282 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff] Apr 24 23:33:16.995377 kernel: pci 0000:00:02.7: bridge window [mem 0x80a00000-0x80bfffff] Apr 24 23:33:16.995472 kernel: pci 0000:00:02.7: bridge window [mem 0xc020000000-0xc03fffffff 64bit pref] Apr 24 23:33:16.995569 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Apr 24 23:33:16.995669 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff] Apr 24 23:33:16.995764 kernel: pci 0000:00:03.0: bridge window [mem 0x80800000-0x809fffff] Apr 24 23:33:16.995859 kernel: pci 0000:00:03.0: bridge window [mem 0xc040000000-0xc05fffffff 64bit pref] Apr 24 23:33:16.998986 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Apr 24 23:33:16.999092 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Apr 24 23:33:16.999186 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Apr 24 23:33:16.999275 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xdfffffff window] Apr 24 23:33:16.999362 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Apr 24 23:33:16.999451 kernel: pci_bus 0000:00: resource 9 [mem 0xc000000000-0xc7ffffffff window] Apr 24 23:33:16.999554 kernel: pci_bus 0000:01: resource 1 [mem 0x81200000-0x812fffff] Apr 24 23:33:16.999649 kernel: pci_bus 0000:01: resource 2 [mem 0xc060000000-0xc0600fffff 64bit pref] Apr 24 23:33:16.999748 kernel: pci_bus 0000:02: resource 1 [mem 0x81100000-0x811fffff] Apr 24 23:33:16.999852 kernel: pci_bus 0000:03: resource 1 [mem 0x81000000-0x810fffff] Apr 24 23:33:16.999958 kernel: pci_bus 0000:03: resource 2 [mem 0xc060100000-0xc0601fffff 64bit pref] Apr 24 23:33:17.000067 kernel: pci_bus 0000:04: resource 2 [mem 0xc060200000-0xc0602fffff 64bit pref] Apr 24 23:33:17.000166 kernel: pci_bus 0000:05: resource 1 [mem 0x80f00000-0x80ffffff] Apr 24 23:33:17.000264 kernel: pci_bus 0000:05: resource 2 [mem 0xc060300000-0xc0603fffff 64bit pref] Apr 24 23:33:17.000393 kernel: pci_bus 0000:06: resource 1 [mem 0x80e00000-0x80efffff] Apr 24 23:33:17.000512 kernel: pci_bus 0000:06: resource 2 [mem 0xc060400000-0xc0604fffff 64bit pref] Apr 24 23:33:17.000616 kernel: pci_bus 0000:07: resource 0 [io 0x1000-0x1fff] Apr 24 23:33:17.000710 kernel: pci_bus 0000:07: resource 1 [mem 0x80c00000-0x80dfffff] Apr 24 23:33:17.000815 kernel: pci_bus 0000:07: resource 2 [mem 0xc000000000-0xc01fffffff 64bit pref] Apr 24 23:33:17.002995 kernel: pci_bus 0000:08: resource 0 [io 0x2000-0x2fff] Apr 24 23:33:17.003106 kernel: pci_bus 0000:08: resource 1 [mem 0x80a00000-0x80bfffff] Apr 24 23:33:17.003206 kernel: pci_bus 0000:08: resource 2 [mem 0xc020000000-0xc03fffffff 64bit pref] Apr 24 23:33:17.003305 kernel: pci_bus 0000:09: resource 0 [io 0x3000-0x3fff] Apr 24 23:33:17.003399 kernel: pci_bus 0000:09: resource 1 [mem 0x80800000-0x809fffff] Apr 24 23:33:17.003491 kernel: pci_bus 0000:09: resource 2 [mem 0xc040000000-0xc05fffffff 64bit pref] Apr 24 23:33:17.003500 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Apr 24 23:33:17.003506 kernel: PCI: CLS 0 bytes, default 64 Apr 24 23:33:17.003512 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Apr 24 23:33:17.003517 kernel: software IO TLB: mapped [mem 0x0000000077ffd000-0x000000007bffd000] (64MB) Apr 24 23:33:17.003526 kernel: Initialise system trusted keyrings Apr 24 23:33:17.003531 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 24 23:33:17.003537 kernel: Key type asymmetric registered Apr 24 23:33:17.003543 kernel: Asymmetric key parser 'x509' registered Apr 24 23:33:17.003549 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Apr 24 23:33:17.003554 kernel: io scheduler mq-deadline registered Apr 24 23:33:17.003560 kernel: io scheduler kyber registered Apr 24 23:33:17.003565 kernel: io scheduler bfq registered Apr 24 23:33:17.003667 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Apr 24 23:33:17.003770 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Apr 24 23:33:17.003867 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Apr 24 23:33:17.003984 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Apr 24 23:33:17.004083 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Apr 24 23:33:17.004179 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Apr 24 23:33:17.004275 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Apr 24 23:33:17.004371 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Apr 24 23:33:17.004466 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Apr 24 23:33:17.004565 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Apr 24 23:33:17.004661 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Apr 24 23:33:17.004757 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Apr 24 23:33:17.004852 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Apr 24 23:33:17.006055 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Apr 24 23:33:17.006180 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Apr 24 23:33:17.006283 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Apr 24 23:33:17.006291 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Apr 24 23:33:17.006389 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 32 Apr 24 23:33:17.006490 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 32 Apr 24 23:33:17.006497 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Apr 24 23:33:17.006503 kernel: ACPI: \_SB_.GSIF: Enabled at IRQ 21 Apr 24 23:33:17.006509 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 24 23:33:17.006515 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Apr 24 23:33:17.006520 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Apr 24 23:33:17.006526 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Apr 24 23:33:17.006531 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Apr 24 23:33:17.006634 kernel: rtc_cmos 00:03: RTC can wake from S4 Apr 24 23:33:17.006645 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Apr 24 23:33:17.006736 kernel: rtc_cmos 00:03: registered as rtc0 Apr 24 23:33:17.006827 kernel: rtc_cmos 00:03: setting system clock to 2026-04-24T23:33:16 UTC (1777073596) Apr 24 23:33:17.006918 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Apr 24 23:33:17.006925 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Apr 24 23:33:17.006931 kernel: efifb: probing for efifb Apr 24 23:33:17.006937 kernel: efifb: framebuffer at 0x80000000, using 4032k, total 4032k Apr 24 23:33:17.009910 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Apr 24 23:33:17.009922 kernel: efifb: scrolling: redraw Apr 24 23:33:17.009928 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 24 23:33:17.009934 kernel: Console: switching to colour frame buffer device 160x50 Apr 24 23:33:17.009940 kernel: fb0: EFI VGA frame buffer device Apr 24 23:33:17.009980 kernel: pstore: Using crash dump compression: deflate Apr 24 23:33:17.009986 kernel: pstore: Registered efi_pstore as persistent store backend Apr 24 23:33:17.009992 kernel: NET: Registered PF_INET6 protocol family Apr 24 23:33:17.009998 kernel: Segment Routing with IPv6 Apr 24 23:33:17.010008 kernel: In-situ OAM (IOAM) with IPv6 Apr 24 23:33:17.010014 kernel: NET: Registered PF_PACKET protocol family Apr 24 23:33:17.010019 kernel: Key type dns_resolver registered Apr 24 23:33:17.010025 kernel: IPI shorthand broadcast: enabled Apr 24 23:33:17.010030 kernel: sched_clock: Marking stable (1334011191, 217369575)->(1600210723, -48829957) Apr 24 23:33:17.010036 kernel: registered taskstats version 1 Apr 24 23:33:17.010042 kernel: Loading compiled-in X.509 certificates Apr 24 23:33:17.010048 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: 507f116e6718ec7535b55c873de10edf9b6fe124' Apr 24 23:33:17.010054 kernel: Key type .fscrypt registered Apr 24 23:33:17.010062 kernel: Key type fscrypt-provisioning registered Apr 24 23:33:17.010067 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 24 23:33:17.010073 kernel: ima: Allocated hash algorithm: sha1 Apr 24 23:33:17.010079 kernel: ima: No architecture policies found Apr 24 23:33:17.010084 kernel: clk: Disabling unused clocks Apr 24 23:33:17.010090 kernel: Freeing unused kernel image (initmem) memory: 42896K Apr 24 23:33:17.010095 kernel: Write protecting the kernel read-only data: 36864k Apr 24 23:33:17.010103 kernel: Freeing unused kernel image (rodata/data gap) memory: 1824K Apr 24 23:33:17.010109 kernel: Run /init as init process Apr 24 23:33:17.010117 kernel: with arguments: Apr 24 23:33:17.010123 kernel: /init Apr 24 23:33:17.010128 kernel: with environment: Apr 24 23:33:17.010134 kernel: HOME=/ Apr 24 23:33:17.010139 kernel: TERM=linux Apr 24 23:33:17.010147 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 24 23:33:17.010155 systemd[1]: Detected virtualization kvm. Apr 24 23:33:17.010162 systemd[1]: Detected architecture x86-64. Apr 24 23:33:17.010170 systemd[1]: Running in initrd. Apr 24 23:33:17.010176 systemd[1]: No hostname configured, using default hostname. Apr 24 23:33:17.010183 systemd[1]: Hostname set to . Apr 24 23:33:17.010193 systemd[1]: Initializing machine ID from VM UUID. Apr 24 23:33:17.010202 systemd[1]: Queued start job for default target initrd.target. Apr 24 23:33:17.010211 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 24 23:33:17.010221 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 24 23:33:17.010231 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 24 23:33:17.010240 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 24 23:33:17.010246 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 24 23:33:17.010252 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 24 23:33:17.010259 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 24 23:33:17.010265 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 24 23:33:17.010271 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 24 23:33:17.010279 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 24 23:33:17.010286 systemd[1]: Reached target paths.target - Path Units. Apr 24 23:33:17.010295 systemd[1]: Reached target slices.target - Slice Units. Apr 24 23:33:17.010304 systemd[1]: Reached target swap.target - Swaps. Apr 24 23:33:17.010314 systemd[1]: Reached target timers.target - Timer Units. Apr 24 23:33:17.010321 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 24 23:33:17.010327 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 24 23:33:17.010333 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 24 23:33:17.010339 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 24 23:33:17.010347 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 24 23:33:17.010353 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 24 23:33:17.010359 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 24 23:33:17.010365 systemd[1]: Reached target sockets.target - Socket Units. Apr 24 23:33:17.010371 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 24 23:33:17.010377 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 24 23:33:17.010387 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 24 23:33:17.010396 systemd[1]: Starting systemd-fsck-usr.service... Apr 24 23:33:17.010405 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 24 23:33:17.010414 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 24 23:33:17.010445 systemd-journald[188]: Collecting audit messages is disabled. Apr 24 23:33:17.010461 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 24 23:33:17.010471 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 24 23:33:17.010481 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 24 23:33:17.010490 systemd[1]: Finished systemd-fsck-usr.service. Apr 24 23:33:17.010499 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 24 23:33:17.010505 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 24 23:33:17.010514 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 24 23:33:17.010521 systemd-journald[188]: Journal started Apr 24 23:33:17.010534 systemd-journald[188]: Runtime Journal (/run/log/journal/80b0573d7d854876b728047d0c9e80a2) is 8.0M, max 76.3M, 68.3M free. Apr 24 23:33:16.989986 systemd-modules-load[189]: Inserted module 'overlay' Apr 24 23:33:17.015007 systemd[1]: Started systemd-journald.service - Journal Service. Apr 24 23:33:17.015318 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 24 23:33:17.021374 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 24 23:33:17.025968 kernel: Bridge firewalling registered Apr 24 23:33:17.025385 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 24 23:33:17.025586 systemd-modules-load[189]: Inserted module 'br_netfilter' Apr 24 23:33:17.027590 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 24 23:33:17.028190 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 24 23:33:17.038078 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 24 23:33:17.039567 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 24 23:33:17.040198 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 24 23:33:17.043090 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 24 23:33:17.050122 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 24 23:33:17.050709 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 24 23:33:17.053100 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 24 23:33:17.059722 dracut-cmdline[219]: dracut-dracut-053 Apr 24 23:33:17.062614 dracut-cmdline[219]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=c8442747465ed99a522e07b8746f6a7817fb39c2025d7438698e3b90e9c0defb Apr 24 23:33:17.085781 systemd-resolved[223]: Positive Trust Anchors: Apr 24 23:33:17.085797 systemd-resolved[223]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 24 23:33:17.085818 systemd-resolved[223]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 24 23:33:17.088037 systemd-resolved[223]: Defaulting to hostname 'linux'. Apr 24 23:33:17.089836 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 24 23:33:17.090306 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 24 23:33:17.127992 kernel: SCSI subsystem initialized Apr 24 23:33:17.136994 kernel: Loading iSCSI transport class v2.0-870. Apr 24 23:33:17.145988 kernel: iscsi: registered transport (tcp) Apr 24 23:33:17.163482 kernel: iscsi: registered transport (qla4xxx) Apr 24 23:33:17.163520 kernel: QLogic iSCSI HBA Driver Apr 24 23:33:17.203421 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 24 23:33:17.208074 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 24 23:33:17.232129 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 24 23:33:17.232169 kernel: device-mapper: uevent: version 1.0.3 Apr 24 23:33:17.232179 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 24 23:33:17.270994 kernel: raid6: avx512x4 gen() 42517 MB/s Apr 24 23:33:17.289000 kernel: raid6: avx512x2 gen() 44950 MB/s Apr 24 23:33:17.307001 kernel: raid6: avx512x1 gen() 42854 MB/s Apr 24 23:33:17.325029 kernel: raid6: avx2x4 gen() 43754 MB/s Apr 24 23:33:17.343037 kernel: raid6: avx2x2 gen() 49006 MB/s Apr 24 23:33:17.362091 kernel: raid6: avx2x1 gen() 37889 MB/s Apr 24 23:33:17.362160 kernel: raid6: using algorithm avx2x2 gen() 49006 MB/s Apr 24 23:33:17.382105 kernel: raid6: .... xor() 35852 MB/s, rmw enabled Apr 24 23:33:17.382182 kernel: raid6: using avx512x2 recovery algorithm Apr 24 23:33:17.399027 kernel: xor: automatically using best checksumming function avx Apr 24 23:33:17.508013 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 24 23:33:17.521854 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 24 23:33:17.526087 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 24 23:33:17.551564 systemd-udevd[406]: Using default interface naming scheme 'v255'. Apr 24 23:33:17.556195 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 24 23:33:17.564109 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 24 23:33:17.575567 dracut-pre-trigger[411]: rd.md=0: removing MD RAID activation Apr 24 23:33:17.605501 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 24 23:33:17.610120 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 24 23:33:17.681398 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 24 23:33:17.692206 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 24 23:33:17.703304 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 24 23:33:17.704196 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 24 23:33:17.704846 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 24 23:33:17.705387 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 24 23:33:17.711108 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 24 23:33:17.721788 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 24 23:33:17.763019 kernel: scsi host0: Virtio SCSI HBA Apr 24 23:33:17.772978 kernel: cryptd: max_cpu_qlen set to 1000 Apr 24 23:33:17.778977 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Apr 24 23:33:17.780976 kernel: libata version 3.00 loaded. Apr 24 23:33:17.790443 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 24 23:33:17.790538 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 24 23:33:17.791031 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 24 23:33:17.791339 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 24 23:33:17.791415 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 24 23:33:17.791741 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 24 23:33:17.797117 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 24 23:33:17.808556 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 24 23:33:17.816570 kernel: ACPI: bus type USB registered Apr 24 23:33:17.816590 kernel: usbcore: registered new interface driver usbfs Apr 24 23:33:17.816599 kernel: usbcore: registered new interface driver hub Apr 24 23:33:17.816615 kernel: usbcore: registered new device driver usb Apr 24 23:33:17.808718 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 24 23:33:17.822228 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 24 23:33:17.830888 kernel: ahci 0000:00:1f.2: version 3.0 Apr 24 23:33:17.834112 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Apr 24 23:33:17.840068 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Apr 24 23:33:17.840229 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Apr 24 23:33:17.852183 kernel: scsi host1: ahci Apr 24 23:33:17.854883 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 24 23:33:17.857711 kernel: scsi host2: ahci Apr 24 23:33:17.857855 kernel: AVX2 version of gcm_enc/dec engaged. Apr 24 23:33:17.857864 kernel: scsi host3: ahci Apr 24 23:33:17.864015 kernel: scsi host4: ahci Apr 24 23:33:17.863225 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 24 23:33:17.871514 kernel: AES CTR mode by8 optimization enabled Apr 24 23:33:17.871554 kernel: scsi host5: ahci Apr 24 23:33:17.873497 kernel: scsi host6: ahci Apr 24 23:33:17.884232 kernel: ata1: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380100 irq 48 Apr 24 23:33:17.884287 kernel: ata2: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380180 irq 48 Apr 24 23:33:17.884297 kernel: ata3: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380200 irq 48 Apr 24 23:33:17.884305 kernel: ata4: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380280 irq 48 Apr 24 23:33:17.884312 kernel: ata5: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380300 irq 48 Apr 24 23:33:17.889783 kernel: ata6: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380380 irq 48 Apr 24 23:33:17.889812 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Apr 24 23:33:17.895033 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Apr 24 23:33:17.895238 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Apr 24 23:33:17.900464 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Apr 24 23:33:17.902122 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Apr 24 23:33:17.902245 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Apr 24 23:33:17.906977 kernel: hub 1-0:1.0: USB hub found Apr 24 23:33:17.909260 kernel: hub 1-0:1.0: 4 ports detected Apr 24 23:33:17.909409 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Apr 24 23:33:17.912172 kernel: hub 2-0:1.0: USB hub found Apr 24 23:33:17.914471 kernel: hub 2-0:1.0: 4 ports detected Apr 24 23:33:17.915301 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 24 23:33:18.156064 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Apr 24 23:33:18.200961 kernel: ata4: SATA link down (SStatus 0 SControl 300) Apr 24 23:33:18.201024 kernel: ata2: SATA link down (SStatus 0 SControl 300) Apr 24 23:33:18.208655 kernel: ata3: SATA link down (SStatus 0 SControl 300) Apr 24 23:33:18.208677 kernel: ata5: SATA link down (SStatus 0 SControl 300) Apr 24 23:33:18.214019 kernel: ata6: SATA link down (SStatus 0 SControl 300) Apr 24 23:33:18.214086 kernel: ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Apr 24 23:33:18.217228 kernel: ata1.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Apr 24 23:33:18.217270 kernel: ata1.00: applying bridge limits Apr 24 23:33:18.219979 kernel: ata1.00: configured for UDMA/100 Apr 24 23:33:18.223022 kernel: scsi 1:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Apr 24 23:33:18.259077 kernel: sd 0:0:0:0: Power-on or device reset occurred Apr 24 23:33:18.266137 kernel: sd 0:0:0:0: [sda] 160006144 512-byte logical blocks: (81.9 GB/76.3 GiB) Apr 24 23:33:18.266308 kernel: sd 0:0:0:0: [sda] Write Protect is off Apr 24 23:33:18.270148 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Apr 24 23:33:18.270307 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Apr 24 23:33:18.284062 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 24 23:33:18.284114 kernel: GPT:17805311 != 160006143 Apr 24 23:33:18.284136 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 24 23:33:18.286151 kernel: GPT:17805311 != 160006143 Apr 24 23:33:18.287531 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 24 23:33:18.289055 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 24 23:33:18.292329 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Apr 24 23:33:18.309171 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 24 23:33:18.309219 kernel: sr 1:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Apr 24 23:33:18.311182 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 24 23:33:18.324740 kernel: usbcore: registered new interface driver usbhid Apr 24 23:33:18.324785 kernel: usbhid: USB HID core driver Apr 24 23:33:18.326980 kernel: sr 1:0:0:0: Attached scsi CD-ROM sr0 Apr 24 23:33:18.343318 kernel: BTRFS: device fsid 077bb4ac-fe88-409a-8f61-fdf28cadf681 devid 1 transid 31 /dev/sda3 scanned by (udev-worker) (457) Apr 24 23:33:18.343355 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input2 Apr 24 23:33:18.347741 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Apr 24 23:33:18.352538 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Apr 24 23:33:18.359689 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Apr 24 23:33:18.359875 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/sda6 scanned by (udev-worker) (469) Apr 24 23:33:18.365648 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Apr 24 23:33:18.366737 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Apr 24 23:33:18.377088 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 24 23:33:18.384695 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Apr 24 23:33:18.385905 disk-uuid[587]: Primary Header is updated. Apr 24 23:33:18.385905 disk-uuid[587]: Secondary Entries is updated. Apr 24 23:33:18.385905 disk-uuid[587]: Secondary Header is updated. Apr 24 23:33:19.400049 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 24 23:33:19.401077 disk-uuid[588]: The operation has completed successfully. Apr 24 23:33:19.452595 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 24 23:33:19.452697 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 24 23:33:19.466071 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 24 23:33:19.470436 sh[599]: Success Apr 24 23:33:19.482999 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Apr 24 23:33:19.526568 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 24 23:33:19.541421 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 24 23:33:19.544646 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 24 23:33:19.561673 kernel: BTRFS info (device dm-0): first mount of filesystem 077bb4ac-fe88-409a-8f61-fdf28cadf681 Apr 24 23:33:19.561701 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Apr 24 23:33:19.561710 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 24 23:33:19.565314 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 24 23:33:19.568834 kernel: BTRFS info (device dm-0): using free space tree Apr 24 23:33:19.577964 kernel: BTRFS info (device dm-0): enabling ssd optimizations Apr 24 23:33:19.579673 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 24 23:33:19.580568 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 24 23:33:19.589139 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 24 23:33:19.592125 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 24 23:33:19.604295 kernel: BTRFS info (device sda6): first mount of filesystem 926930fb-88b5-4cf4-bdd1-3374ab036b7b Apr 24 23:33:19.604340 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Apr 24 23:33:19.607975 kernel: BTRFS info (device sda6): using free space tree Apr 24 23:33:19.615211 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 24 23:33:19.615263 kernel: BTRFS info (device sda6): auto enabling async discard Apr 24 23:33:19.625705 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 24 23:33:19.628880 kernel: BTRFS info (device sda6): last unmount of filesystem 926930fb-88b5-4cf4-bdd1-3374ab036b7b Apr 24 23:33:19.634722 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 24 23:33:19.640899 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 24 23:33:19.695644 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 24 23:33:19.703244 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 24 23:33:19.712157 ignition[719]: Ignition 2.19.0 Apr 24 23:33:19.712164 ignition[719]: Stage: fetch-offline Apr 24 23:33:19.712195 ignition[719]: no configs at "/usr/lib/ignition/base.d" Apr 24 23:33:19.712204 ignition[719]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 24 23:33:19.712307 ignition[719]: parsed url from cmdline: "" Apr 24 23:33:19.715676 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 24 23:33:19.712311 ignition[719]: no config URL provided Apr 24 23:33:19.712315 ignition[719]: reading system config file "/usr/lib/ignition/user.ign" Apr 24 23:33:19.712323 ignition[719]: no config at "/usr/lib/ignition/user.ign" Apr 24 23:33:19.712328 ignition[719]: failed to fetch config: resource requires networking Apr 24 23:33:19.712450 ignition[719]: Ignition finished successfully Apr 24 23:33:19.720780 systemd-networkd[783]: lo: Link UP Apr 24 23:33:19.720790 systemd-networkd[783]: lo: Gained carrier Apr 24 23:33:19.723185 systemd-networkd[783]: Enumeration completed Apr 24 23:33:19.723678 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 24 23:33:19.723998 systemd-networkd[783]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:19.724002 systemd-networkd[783]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 24 23:33:19.724662 systemd-networkd[783]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:19.724667 systemd-networkd[783]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 24 23:33:19.725279 systemd-networkd[783]: eth0: Link UP Apr 24 23:33:19.725284 systemd-networkd[783]: eth0: Gained carrier Apr 24 23:33:19.725290 systemd-networkd[783]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:19.726058 systemd[1]: Reached target network.target - Network. Apr 24 23:33:19.729336 systemd-networkd[783]: eth1: Link UP Apr 24 23:33:19.729340 systemd-networkd[783]: eth1: Gained carrier Apr 24 23:33:19.729348 systemd-networkd[783]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:19.731098 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Apr 24 23:33:19.743224 ignition[788]: Ignition 2.19.0 Apr 24 23:33:19.743825 ignition[788]: Stage: fetch Apr 24 23:33:19.743992 ignition[788]: no configs at "/usr/lib/ignition/base.d" Apr 24 23:33:19.744004 ignition[788]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 24 23:33:19.744100 ignition[788]: parsed url from cmdline: "" Apr 24 23:33:19.744104 ignition[788]: no config URL provided Apr 24 23:33:19.744109 ignition[788]: reading system config file "/usr/lib/ignition/user.ign" Apr 24 23:33:19.744117 ignition[788]: no config at "/usr/lib/ignition/user.ign" Apr 24 23:33:19.744132 ignition[788]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Apr 24 23:33:19.744280 ignition[788]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Apr 24 23:33:19.764996 systemd-networkd[783]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Apr 24 23:33:19.770025 systemd-networkd[783]: eth0: DHCPv4 address 95.217.179.164/32, gateway 172.31.1.1 acquired from 172.31.1.1 Apr 24 23:33:19.944464 ignition[788]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Apr 24 23:33:19.950201 ignition[788]: GET result: OK Apr 24 23:33:19.950748 ignition[788]: parsing config with SHA512: 4863058405541f35fe3cbe2173cacd0e9f40c5ad39b663b9f7609e4f44f0002dac0090ec0f08b6694493b52cc3fb2081a7e30d3792dec58a847ef03e80b5e3ae Apr 24 23:33:19.955939 unknown[788]: fetched base config from "system" Apr 24 23:33:19.955991 unknown[788]: fetched base config from "system" Apr 24 23:33:19.956674 ignition[788]: fetch: fetch complete Apr 24 23:33:19.956003 unknown[788]: fetched user config from "hetzner" Apr 24 23:33:19.956685 ignition[788]: fetch: fetch passed Apr 24 23:33:19.956757 ignition[788]: Ignition finished successfully Apr 24 23:33:19.962517 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Apr 24 23:33:19.969291 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 24 23:33:20.006388 ignition[795]: Ignition 2.19.0 Apr 24 23:33:20.006417 ignition[795]: Stage: kargs Apr 24 23:33:20.006744 ignition[795]: no configs at "/usr/lib/ignition/base.d" Apr 24 23:33:20.006768 ignition[795]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 24 23:33:20.008218 ignition[795]: kargs: kargs passed Apr 24 23:33:20.008302 ignition[795]: Ignition finished successfully Apr 24 23:33:20.011051 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 24 23:33:20.018250 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 24 23:33:20.041693 ignition[802]: Ignition 2.19.0 Apr 24 23:33:20.041711 ignition[802]: Stage: disks Apr 24 23:33:20.042127 ignition[802]: no configs at "/usr/lib/ignition/base.d" Apr 24 23:33:20.042152 ignition[802]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 24 23:33:20.045559 ignition[802]: disks: disks passed Apr 24 23:33:20.045659 ignition[802]: Ignition finished successfully Apr 24 23:33:20.050104 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 24 23:33:20.051436 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 24 23:33:20.052352 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 24 23:33:20.053566 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 24 23:33:20.054769 systemd[1]: Reached target sysinit.target - System Initialization. Apr 24 23:33:20.056071 systemd[1]: Reached target basic.target - Basic System. Apr 24 23:33:20.063231 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 24 23:33:20.087749 systemd-fsck[810]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Apr 24 23:33:20.091748 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 24 23:33:20.099135 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 24 23:33:20.191964 kernel: EXT4-fs (sda9): mounted filesystem ae73d4a7-3ef8-4c50-8348-4aeb952085ba r/w with ordered data mode. Quota mode: none. Apr 24 23:33:20.193124 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 24 23:33:20.195266 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 24 23:33:20.202151 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 24 23:33:20.205109 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 24 23:33:20.209007 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Apr 24 23:33:20.209682 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 24 23:33:20.209704 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 24 23:33:20.218185 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by mount (818) Apr 24 23:33:20.218247 kernel: BTRFS info (device sda6): first mount of filesystem 926930fb-88b5-4cf4-bdd1-3374ab036b7b Apr 24 23:33:20.218972 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Apr 24 23:33:20.219000 kernel: BTRFS info (device sda6): using free space tree Apr 24 23:33:20.228996 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 24 23:33:20.229070 kernel: BTRFS info (device sda6): auto enabling async discard Apr 24 23:33:20.229899 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 24 23:33:20.233820 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 24 23:33:20.246521 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 24 23:33:20.294238 coreos-metadata[820]: Apr 24 23:33:20.294 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Apr 24 23:33:20.295365 coreos-metadata[820]: Apr 24 23:33:20.295 INFO Fetch successful Apr 24 23:33:20.296759 coreos-metadata[820]: Apr 24 23:33:20.296 INFO wrote hostname ci-4081-3-6-n-e88c017743 to /sysroot/etc/hostname Apr 24 23:33:20.298365 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 24 23:33:20.300074 initrd-setup-root[846]: cut: /sysroot/etc/passwd: No such file or directory Apr 24 23:33:20.303932 initrd-setup-root[854]: cut: /sysroot/etc/group: No such file or directory Apr 24 23:33:20.309272 initrd-setup-root[861]: cut: /sysroot/etc/shadow: No such file or directory Apr 24 23:33:20.313264 initrd-setup-root[868]: cut: /sysroot/etc/gshadow: No such file or directory Apr 24 23:33:20.395313 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 24 23:33:20.398053 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 24 23:33:20.401076 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 24 23:33:20.412981 kernel: BTRFS info (device sda6): last unmount of filesystem 926930fb-88b5-4cf4-bdd1-3374ab036b7b Apr 24 23:33:20.425061 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 24 23:33:20.433156 ignition[941]: INFO : Ignition 2.19.0 Apr 24 23:33:20.433156 ignition[941]: INFO : Stage: mount Apr 24 23:33:20.434157 ignition[941]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 24 23:33:20.434157 ignition[941]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 24 23:33:20.434850 ignition[941]: INFO : mount: mount passed Apr 24 23:33:20.434850 ignition[941]: INFO : Ignition finished successfully Apr 24 23:33:20.435712 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 24 23:33:20.440022 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 24 23:33:20.556813 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 24 23:33:20.563079 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 24 23:33:20.590046 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (951) Apr 24 23:33:20.600902 kernel: BTRFS info (device sda6): first mount of filesystem 926930fb-88b5-4cf4-bdd1-3374ab036b7b Apr 24 23:33:20.600979 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Apr 24 23:33:20.601002 kernel: BTRFS info (device sda6): using free space tree Apr 24 23:33:20.613441 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 24 23:33:20.613511 kernel: BTRFS info (device sda6): auto enabling async discard Apr 24 23:33:20.622316 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 24 23:33:20.654476 ignition[967]: INFO : Ignition 2.19.0 Apr 24 23:33:20.654476 ignition[967]: INFO : Stage: files Apr 24 23:33:20.656112 ignition[967]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 24 23:33:20.656112 ignition[967]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 24 23:33:20.657556 ignition[967]: DEBUG : files: compiled without relabeling support, skipping Apr 24 23:33:20.658664 ignition[967]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 24 23:33:20.658664 ignition[967]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 24 23:33:20.663023 ignition[967]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 24 23:33:20.664020 ignition[967]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 24 23:33:20.664642 ignition[967]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 24 23:33:20.664510 unknown[967]: wrote ssh authorized keys file for user: core Apr 24 23:33:20.667193 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Apr 24 23:33:20.667801 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Apr 24 23:33:20.667801 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 24 23:33:20.667801 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Apr 24 23:33:21.059221 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Apr 24 23:33:21.357361 systemd-networkd[783]: eth0: Gained IPv6LL Apr 24 23:33:21.486074 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 24 23:33:21.488240 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 24 23:33:21.501257 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-x86-64.raw: attempt #1 Apr 24 23:33:21.741414 systemd-networkd[783]: eth1: Gained IPv6LL Apr 24 23:33:21.859294 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Apr 24 23:33:22.223938 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 24 23:33:22.223938 ignition[967]: INFO : files: op(c): [started] processing unit "containerd.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(c): [finished] processing unit "containerd.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(10): [started] processing unit "coreos-metadata.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(10): op(11): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(10): op(11): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(10): [finished] processing unit "coreos-metadata.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Apr 24 23:33:22.228366 ignition[967]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 24 23:33:22.228366 ignition[967]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 24 23:33:22.228366 ignition[967]: INFO : files: files passed Apr 24 23:33:22.228366 ignition[967]: INFO : Ignition finished successfully Apr 24 23:33:22.228136 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 24 23:33:22.236576 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 24 23:33:22.239060 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 24 23:33:22.244019 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 24 23:33:22.244110 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 24 23:33:22.252081 initrd-setup-root-after-ignition[996]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 24 23:33:22.253263 initrd-setup-root-after-ignition[996]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 24 23:33:22.253678 initrd-setup-root-after-ignition[1000]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 24 23:33:22.255111 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 24 23:33:22.256345 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 24 23:33:22.261069 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 24 23:33:22.280012 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 24 23:33:22.280116 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 24 23:33:22.280712 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 24 23:33:22.281364 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 24 23:33:22.282211 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 24 23:33:22.284073 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 24 23:33:22.303457 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 24 23:33:22.309087 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 24 23:33:22.316257 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 24 23:33:22.316702 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 24 23:33:22.317152 systemd[1]: Stopped target timers.target - Timer Units. Apr 24 23:33:22.317555 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 24 23:33:22.317629 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 24 23:33:22.318500 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 24 23:33:22.319318 systemd[1]: Stopped target basic.target - Basic System. Apr 24 23:33:22.320095 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 24 23:33:22.320893 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 24 23:33:22.321585 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 24 23:33:22.322300 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 24 23:33:22.323022 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 24 23:33:22.323766 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 24 23:33:22.324484 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 24 23:33:22.325190 systemd[1]: Stopped target swap.target - Swaps. Apr 24 23:33:22.325871 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 24 23:33:22.325961 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 24 23:33:22.326970 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 24 23:33:22.328009 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 24 23:33:22.328642 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 24 23:33:22.328723 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 24 23:33:22.329442 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 24 23:33:22.329532 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 24 23:33:22.330483 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 24 23:33:22.330565 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 24 23:33:22.331269 systemd[1]: ignition-files.service: Deactivated successfully. Apr 24 23:33:22.331337 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 24 23:33:22.331917 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 24 23:33:22.332002 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 24 23:33:22.342354 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 24 23:33:22.343107 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 24 23:33:22.343529 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 24 23:33:22.347105 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 24 23:33:22.347829 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 24 23:33:22.348302 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 24 23:33:22.349159 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 24 23:33:22.349256 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 24 23:33:22.352973 ignition[1020]: INFO : Ignition 2.19.0 Apr 24 23:33:22.352973 ignition[1020]: INFO : Stage: umount Apr 24 23:33:22.352973 ignition[1020]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 24 23:33:22.352973 ignition[1020]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 24 23:33:22.353030 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 24 23:33:22.356538 ignition[1020]: INFO : umount: umount passed Apr 24 23:33:22.356538 ignition[1020]: INFO : Ignition finished successfully Apr 24 23:33:22.353127 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 24 23:33:22.357651 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 24 23:33:22.357760 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 24 23:33:22.359704 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 24 23:33:22.359743 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 24 23:33:22.360728 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 24 23:33:22.360767 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 24 23:33:22.363013 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 24 23:33:22.363095 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Apr 24 23:33:22.363781 systemd[1]: Stopped target network.target - Network. Apr 24 23:33:22.364116 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 24 23:33:22.364155 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 24 23:33:22.364483 systemd[1]: Stopped target paths.target - Path Units. Apr 24 23:33:22.364779 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 24 23:33:22.369029 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 24 23:33:22.369742 systemd[1]: Stopped target slices.target - Slice Units. Apr 24 23:33:22.370373 systemd[1]: Stopped target sockets.target - Socket Units. Apr 24 23:33:22.370727 systemd[1]: iscsid.socket: Deactivated successfully. Apr 24 23:33:22.370770 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 24 23:33:22.371163 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 24 23:33:22.371201 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 24 23:33:22.371825 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 24 23:33:22.371864 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 24 23:33:22.372541 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 24 23:33:22.372588 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 24 23:33:22.373368 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 24 23:33:22.373979 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 24 23:33:22.375305 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 24 23:33:22.378031 systemd-networkd[783]: eth0: DHCPv6 lease lost Apr 24 23:33:22.382099 systemd-networkd[783]: eth1: DHCPv6 lease lost Apr 24 23:33:22.384199 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 24 23:33:22.384623 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 24 23:33:22.385646 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 24 23:33:22.386469 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 24 23:33:22.387751 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 24 23:33:22.387806 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 24 23:33:22.396012 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 24 23:33:22.396319 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 24 23:33:22.396360 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 24 23:33:22.396696 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 24 23:33:22.396729 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 24 23:33:22.397343 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 24 23:33:22.397385 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 24 23:33:22.397724 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 24 23:33:22.397756 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 24 23:33:22.398158 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 24 23:33:22.402816 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 24 23:33:22.402898 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 24 23:33:22.405590 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 24 23:33:22.405659 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 24 23:33:22.413871 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 24 23:33:22.414348 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 24 23:33:22.417258 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 24 23:33:22.417390 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 24 23:33:22.418395 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 24 23:33:22.418451 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 24 23:33:22.418809 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 24 23:33:22.418841 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 24 23:33:22.419207 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 24 23:33:22.419244 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 24 23:33:22.419645 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 24 23:33:22.419677 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 24 23:33:22.420375 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 24 23:33:22.420413 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 24 23:33:22.426082 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 24 23:33:22.426404 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 24 23:33:22.426446 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 24 23:33:22.426795 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 24 23:33:22.426827 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 24 23:33:22.431506 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 24 23:33:22.432021 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 24 23:33:22.433155 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 24 23:33:22.434612 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 24 23:33:22.444599 systemd[1]: Switching root. Apr 24 23:33:22.483129 systemd-journald[188]: Journal stopped Apr 24 23:33:23.563612 systemd-journald[188]: Received SIGTERM from PID 1 (systemd). Apr 24 23:33:23.563675 kernel: SELinux: policy capability network_peer_controls=1 Apr 24 23:33:23.563689 kernel: SELinux: policy capability open_perms=1 Apr 24 23:33:23.563697 kernel: SELinux: policy capability extended_socket_class=1 Apr 24 23:33:23.563705 kernel: SELinux: policy capability always_check_network=0 Apr 24 23:33:23.563713 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 24 23:33:23.563726 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 24 23:33:23.563736 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 24 23:33:23.563745 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 24 23:33:23.563761 kernel: audit: type=1403 audit(1777073602.685:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 24 23:33:23.563776 systemd[1]: Successfully loaded SELinux policy in 58.625ms. Apr 24 23:33:23.563797 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 9.650ms. Apr 24 23:33:23.563807 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 24 23:33:23.563818 systemd[1]: Detected virtualization kvm. Apr 24 23:33:23.563827 systemd[1]: Detected architecture x86-64. Apr 24 23:33:23.563836 systemd[1]: Detected first boot. Apr 24 23:33:23.563845 systemd[1]: Hostname set to . Apr 24 23:33:23.563857 systemd[1]: Initializing machine ID from VM UUID. Apr 24 23:33:23.563865 zram_generator::config[1081]: No configuration found. Apr 24 23:33:23.563875 systemd[1]: Populated /etc with preset unit settings. Apr 24 23:33:23.563884 systemd[1]: Queued start job for default target multi-user.target. Apr 24 23:33:23.563896 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Apr 24 23:33:23.563905 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 24 23:33:23.563914 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 24 23:33:23.563922 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 24 23:33:23.563931 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 24 23:33:23.563940 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 24 23:33:23.565984 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 24 23:33:23.566000 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 24 23:33:23.566014 systemd[1]: Created slice user.slice - User and Session Slice. Apr 24 23:33:23.566023 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 24 23:33:23.566032 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 24 23:33:23.566041 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 24 23:33:23.566050 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 24 23:33:23.566059 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 24 23:33:23.566069 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 24 23:33:23.566087 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Apr 24 23:33:23.566097 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 24 23:33:23.566108 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 24 23:33:23.566118 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 24 23:33:23.566132 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 24 23:33:23.566141 systemd[1]: Reached target slices.target - Slice Units. Apr 24 23:33:23.566150 systemd[1]: Reached target swap.target - Swaps. Apr 24 23:33:23.566158 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 24 23:33:23.566167 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 24 23:33:23.566179 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 24 23:33:23.566187 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 24 23:33:23.566196 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 24 23:33:23.566205 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 24 23:33:23.566214 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 24 23:33:23.566223 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 24 23:33:23.566232 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 24 23:33:23.566241 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 24 23:33:23.566250 systemd[1]: Mounting media.mount - External Media Directory... Apr 24 23:33:23.566265 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:23.566277 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 24 23:33:23.566290 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 24 23:33:23.566302 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 24 23:33:23.566314 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 24 23:33:23.566327 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 24 23:33:23.566337 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 24 23:33:23.566346 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 24 23:33:23.566356 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 24 23:33:23.566368 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 24 23:33:23.566377 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 24 23:33:23.566386 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 24 23:33:23.566395 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 24 23:33:23.566404 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 24 23:33:23.566413 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Apr 24 23:33:23.566425 kernel: fuse: init (API version 7.39) Apr 24 23:33:23.566434 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Apr 24 23:33:23.566445 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 24 23:33:23.566454 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 24 23:33:23.566470 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 24 23:33:23.566481 kernel: ACPI: bus type drm_connector registered Apr 24 23:33:23.566490 kernel: loop: module loaded Apr 24 23:33:23.566499 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 24 23:33:23.566530 systemd-journald[1177]: Collecting audit messages is disabled. Apr 24 23:33:23.566554 systemd-journald[1177]: Journal started Apr 24 23:33:23.566570 systemd-journald[1177]: Runtime Journal (/run/log/journal/80b0573d7d854876b728047d0c9e80a2) is 8.0M, max 76.3M, 68.3M free. Apr 24 23:33:23.576444 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 24 23:33:23.576502 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:23.592985 systemd[1]: Started systemd-journald.service - Journal Service. Apr 24 23:33:23.594659 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 24 23:33:23.595385 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 24 23:33:23.597054 systemd[1]: Mounted media.mount - External Media Directory. Apr 24 23:33:23.597704 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 24 23:33:23.598206 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 24 23:33:23.598668 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 24 23:33:23.599556 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 24 23:33:23.600276 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 24 23:33:23.601045 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 24 23:33:23.601272 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 24 23:33:23.602016 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 24 23:33:23.602228 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 24 23:33:23.602992 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 24 23:33:23.603208 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 24 23:33:23.603894 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 24 23:33:23.604116 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 24 23:33:23.604811 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 24 23:33:23.605340 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 24 23:33:23.606097 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 24 23:33:23.606341 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 24 23:33:23.607229 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 24 23:33:23.607964 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 24 23:33:23.608682 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 24 23:33:23.620699 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 24 23:33:23.628108 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 24 23:33:23.632054 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 24 23:33:23.632502 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 24 23:33:23.642201 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 24 23:33:23.647117 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 24 23:33:23.650013 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 24 23:33:23.659735 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 24 23:33:23.660848 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 24 23:33:23.664172 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 24 23:33:23.676058 systemd-journald[1177]: Time spent on flushing to /var/log/journal/80b0573d7d854876b728047d0c9e80a2 is 26.555ms for 1161 entries. Apr 24 23:33:23.676058 systemd-journald[1177]: System Journal (/var/log/journal/80b0573d7d854876b728047d0c9e80a2) is 8.0M, max 584.8M, 576.8M free. Apr 24 23:33:23.714203 systemd-journald[1177]: Received client request to flush runtime journal. Apr 24 23:33:23.670061 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 24 23:33:23.673849 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 24 23:33:23.678041 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 24 23:33:23.687578 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 24 23:33:23.688087 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 24 23:33:23.716617 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 24 23:33:23.721685 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 24 23:33:23.732687 systemd-tmpfiles[1224]: ACLs are not supported, ignoring. Apr 24 23:33:23.732707 systemd-tmpfiles[1224]: ACLs are not supported, ignoring. Apr 24 23:33:23.744385 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 24 23:33:23.750096 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 24 23:33:23.779427 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 24 23:33:23.794133 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 24 23:33:23.806340 udevadm[1243]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Apr 24 23:33:23.814194 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 24 23:33:23.821155 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 24 23:33:23.836220 systemd-tmpfiles[1247]: ACLs are not supported, ignoring. Apr 24 23:33:23.836451 systemd-tmpfiles[1247]: ACLs are not supported, ignoring. Apr 24 23:33:23.843307 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 24 23:33:24.032120 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 24 23:33:24.041261 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 24 23:33:24.063078 systemd-udevd[1253]: Using default interface naming scheme 'v255'. Apr 24 23:33:24.085990 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 24 23:33:24.100191 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 24 23:33:24.132126 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 24 23:33:24.142769 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Apr 24 23:33:24.198728 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 24 23:33:24.205148 kernel: mousedev: PS/2 mouse device common for all mice Apr 24 23:33:24.248966 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 31 scanned by (udev-worker) (1271) Apr 24 23:33:24.266029 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Apr 24 23:33:24.278975 kernel: ACPI: button: Power Button [PWRF] Apr 24 23:33:24.291064 systemd-networkd[1262]: lo: Link UP Apr 24 23:33:24.291375 systemd-networkd[1262]: lo: Gained carrier Apr 24 23:33:24.293766 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Apr 24 23:33:24.297451 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:24.297592 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 24 23:33:24.304908 systemd-networkd[1262]: Enumeration completed Apr 24 23:33:24.305103 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 24 23:33:24.306816 systemd-networkd[1262]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:24.307037 systemd-networkd[1262]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 24 23:33:24.308028 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 24 23:33:24.308762 systemd-networkd[1262]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:24.308801 systemd-networkd[1262]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 24 23:33:24.310565 systemd-networkd[1262]: eth0: Link UP Apr 24 23:33:24.312018 systemd-networkd[1262]: eth0: Gained carrier Apr 24 23:33:24.312077 systemd-networkd[1262]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:24.315186 systemd-networkd[1262]: eth1: Link UP Apr 24 23:33:24.315250 systemd-networkd[1262]: eth1: Gained carrier Apr 24 23:33:24.315284 systemd-networkd[1262]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:24.320366 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 24 23:33:24.321997 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 24 23:33:24.322026 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 24 23:33:24.322059 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:24.322147 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 24 23:33:24.322797 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 24 23:33:24.323274 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 24 23:33:24.331986 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 24 23:33:24.332158 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 24 23:33:24.335712 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 24 23:33:24.336280 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 24 23:33:24.353046 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 24 23:33:24.353208 systemd-networkd[1262]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Apr 24 23:33:24.353438 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 24 23:33:24.353478 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 24 23:33:24.356707 systemd-networkd[1262]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:24.371254 systemd-networkd[1262]: eth0: DHCPv4 address 95.217.179.164/32, gateway 172.31.1.1 acquired from 172.31.1.1 Apr 24 23:33:24.371415 systemd-networkd[1262]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 24 23:33:24.376991 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Apr 24 23:33:24.394005 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Apr 24 23:33:24.394241 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Apr 24 23:33:24.394374 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Apr 24 23:33:24.394518 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Apr 24 23:33:24.401966 kernel: [drm] pci: virtio-vga detected at 0000:00:01.0 Apr 24 23:33:24.409460 kernel: Console: switching to colour dummy device 80x25 Apr 24 23:33:24.411029 kernel: virtio-pci 0000:00:01.0: vgaarb: deactivate vga console Apr 24 23:33:24.411262 kernel: EDAC MC: Ver: 3.0.0 Apr 24 23:33:24.418777 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Apr 24 23:33:24.418804 kernel: [drm] features: -context_init Apr 24 23:33:24.418816 kernel: [drm] number of scanouts: 1 Apr 24 23:33:24.420563 kernel: [drm] number of cap sets: 0 Apr 24 23:33:24.421232 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 24 23:33:24.423964 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Apr 24 23:33:24.430276 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Apr 24 23:33:24.430301 kernel: Console: switching to colour frame buffer device 160x50 Apr 24 23:33:24.425406 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 24 23:33:24.427094 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 24 23:33:24.437970 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Apr 24 23:33:24.447249 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 24 23:33:24.449274 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 24 23:33:24.449560 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 24 23:33:24.453170 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 24 23:33:24.505689 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 24 23:33:24.567752 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Apr 24 23:33:24.575206 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Apr 24 23:33:24.593600 lvm[1326]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 24 23:33:24.630684 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Apr 24 23:33:24.631881 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 24 23:33:24.644274 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Apr 24 23:33:24.654364 lvm[1329]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 24 23:33:24.696476 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Apr 24 23:33:24.698742 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 24 23:33:24.698927 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 24 23:33:24.698981 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 24 23:33:24.699047 systemd[1]: Reached target machines.target - Containers. Apr 24 23:33:24.700035 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Apr 24 23:33:24.707057 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 24 23:33:24.708510 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 24 23:33:24.708804 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 24 23:33:24.714395 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 24 23:33:24.720120 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Apr 24 23:33:24.736272 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 24 23:33:24.739879 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 24 23:33:24.759865 kernel: loop0: detected capacity change from 0 to 228704 Apr 24 23:33:24.760734 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 24 23:33:24.762315 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 24 23:33:24.766298 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Apr 24 23:33:24.791437 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 24 23:33:24.812974 kernel: loop1: detected capacity change from 0 to 142488 Apr 24 23:33:24.851978 kernel: loop2: detected capacity change from 0 to 140768 Apr 24 23:33:24.891004 kernel: loop3: detected capacity change from 0 to 8 Apr 24 23:33:24.907991 kernel: loop4: detected capacity change from 0 to 228704 Apr 24 23:33:24.930154 kernel: loop5: detected capacity change from 0 to 142488 Apr 24 23:33:24.946983 kernel: loop6: detected capacity change from 0 to 140768 Apr 24 23:33:24.967995 kernel: loop7: detected capacity change from 0 to 8 Apr 24 23:33:24.970261 (sd-merge)[1350]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Apr 24 23:33:24.971940 (sd-merge)[1350]: Merged extensions into '/usr'. Apr 24 23:33:24.986253 systemd[1]: Reloading requested from client PID 1337 ('systemd-sysext') (unit systemd-sysext.service)... Apr 24 23:33:24.986268 systemd[1]: Reloading... Apr 24 23:33:25.045092 zram_generator::config[1378]: No configuration found. Apr 24 23:33:25.100935 ldconfig[1333]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 24 23:33:25.149532 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 24 23:33:25.202158 systemd[1]: Reloading finished in 215 ms. Apr 24 23:33:25.218139 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 24 23:33:25.224828 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 24 23:33:25.234353 systemd[1]: Starting ensure-sysext.service... Apr 24 23:33:25.250269 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 24 23:33:25.259540 systemd[1]: Reloading requested from client PID 1428 ('systemctl') (unit ensure-sysext.service)... Apr 24 23:33:25.259712 systemd[1]: Reloading... Apr 24 23:33:25.263590 systemd-tmpfiles[1429]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 24 23:33:25.263874 systemd-tmpfiles[1429]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 24 23:33:25.264669 systemd-tmpfiles[1429]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 24 23:33:25.264873 systemd-tmpfiles[1429]: ACLs are not supported, ignoring. Apr 24 23:33:25.264935 systemd-tmpfiles[1429]: ACLs are not supported, ignoring. Apr 24 23:33:25.268696 systemd-tmpfiles[1429]: Detected autofs mount point /boot during canonicalization of boot. Apr 24 23:33:25.268774 systemd-tmpfiles[1429]: Skipping /boot Apr 24 23:33:25.278921 systemd-tmpfiles[1429]: Detected autofs mount point /boot during canonicalization of boot. Apr 24 23:33:25.279009 systemd-tmpfiles[1429]: Skipping /boot Apr 24 23:33:25.327980 zram_generator::config[1462]: No configuration found. Apr 24 23:33:25.427342 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 24 23:33:25.486129 systemd[1]: Reloading finished in 225 ms. Apr 24 23:33:25.502495 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 24 23:33:25.515267 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 24 23:33:25.522061 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 24 23:33:25.531018 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 24 23:33:25.536756 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 24 23:33:25.541678 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 24 23:33:25.553375 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:25.553522 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 24 23:33:25.558365 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 24 23:33:25.568383 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 24 23:33:25.576902 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 24 23:33:25.581828 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 24 23:33:25.583028 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:25.583866 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 24 23:33:25.586994 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 24 23:33:25.587178 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 24 23:33:25.590097 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 24 23:33:25.590565 augenrules[1535]: No rules Apr 24 23:33:25.592194 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 24 23:33:25.593528 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 24 23:33:25.600545 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 24 23:33:25.601981 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 24 23:33:25.614145 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:25.614321 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 24 23:33:25.616313 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 24 23:33:25.627165 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 24 23:33:25.633429 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 24 23:33:25.637974 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 24 23:33:25.649217 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 24 23:33:25.650691 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:25.655465 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 24 23:33:25.656660 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 24 23:33:25.656845 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 24 23:33:25.657570 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 24 23:33:25.659666 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 24 23:33:25.665672 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 24 23:33:25.665847 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 24 23:33:25.680126 systemd[1]: Finished ensure-sysext.service. Apr 24 23:33:25.680873 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 24 23:33:25.692275 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 24 23:33:25.694789 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:25.695423 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 24 23:33:25.700082 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 24 23:33:25.701540 systemd-resolved[1519]: Positive Trust Anchors: Apr 24 23:33:25.701781 systemd-resolved[1519]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 24 23:33:25.701831 systemd-resolved[1519]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 24 23:33:25.705376 systemd-resolved[1519]: Using system hostname 'ci-4081-3-6-n-e88c017743'. Apr 24 23:33:25.713081 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 24 23:33:25.718156 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 24 23:33:25.740102 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 24 23:33:25.741369 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 24 23:33:25.745394 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 24 23:33:25.748536 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 24 23:33:25.748569 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 24 23:33:25.748817 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 24 23:33:25.752541 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 24 23:33:25.752726 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 24 23:33:25.753368 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 24 23:33:25.753527 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 24 23:33:25.754148 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 24 23:33:25.754309 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 24 23:33:25.754880 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 24 23:33:25.757102 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 24 23:33:25.760437 systemd[1]: Reached target network.target - Network. Apr 24 23:33:25.762092 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 24 23:33:25.762433 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 24 23:33:25.762493 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 24 23:33:25.808290 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 24 23:33:25.809010 systemd[1]: Reached target sysinit.target - System Initialization. Apr 24 23:33:25.809441 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 24 23:33:25.809777 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 24 23:33:25.811795 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 24 23:33:25.812165 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 24 23:33:25.812189 systemd[1]: Reached target paths.target - Path Units. Apr 24 23:33:25.812492 systemd[1]: Reached target time-set.target - System Time Set. Apr 24 23:33:25.812973 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 24 23:33:25.813391 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 24 23:33:25.813695 systemd[1]: Reached target timers.target - Timer Units. Apr 24 23:33:25.818208 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 24 23:33:25.819900 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 24 23:33:25.824396 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 24 23:33:25.827127 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 24 23:33:25.827467 systemd[1]: Reached target sockets.target - Socket Units. Apr 24 23:33:25.827754 systemd[1]: Reached target basic.target - Basic System. Apr 24 23:33:25.828239 systemd[1]: System is tainted: cgroupsv1 Apr 24 23:33:25.828275 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 24 23:33:25.828294 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 24 23:33:25.829801 systemd[1]: Starting containerd.service - containerd container runtime... Apr 24 23:33:25.835279 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Apr 24 23:33:25.839142 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 24 23:33:26.677471 systemd-resolved[1519]: Clock change detected. Flushing caches. Apr 24 23:33:26.677563 systemd-timesyncd[1575]: Contacted time server 85.121.54.197:123 (0.flatcar.pool.ntp.org). Apr 24 23:33:26.677609 systemd-timesyncd[1575]: Initial clock synchronization to Fri 2026-04-24 23:33:26.677437 UTC. Apr 24 23:33:26.679217 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 24 23:33:26.691278 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 24 23:33:26.691626 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 24 23:33:26.694049 coreos-metadata[1587]: Apr 24 23:33:26.693 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Apr 24 23:33:26.696283 coreos-metadata[1587]: Apr 24 23:33:26.694 INFO Fetch successful Apr 24 23:33:26.696283 coreos-metadata[1587]: Apr 24 23:33:26.694 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Apr 24 23:33:26.696283 coreos-metadata[1587]: Apr 24 23:33:26.696 INFO Fetch successful Apr 24 23:33:26.696627 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 24 23:33:26.704540 jq[1590]: false Apr 24 23:33:26.705240 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 24 23:33:26.715258 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Apr 24 23:33:26.720907 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 24 23:33:26.735255 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 24 23:33:26.744239 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 24 23:33:26.748632 extend-filesystems[1591]: Found loop4 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found loop5 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found loop6 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found loop7 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found sda Apr 24 23:33:26.748632 extend-filesystems[1591]: Found sda1 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found sda2 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found sda3 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found usr Apr 24 23:33:26.748632 extend-filesystems[1591]: Found sda4 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found sda6 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found sda7 Apr 24 23:33:26.748632 extend-filesystems[1591]: Found sda9 Apr 24 23:33:26.748632 extend-filesystems[1591]: Checking size of /dev/sda9 Apr 24 23:33:26.797956 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 19393531 blocks Apr 24 23:33:26.747379 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 24 23:33:26.759122 dbus-daemon[1589]: [system] SELinux support is enabled Apr 24 23:33:26.798794 extend-filesystems[1591]: Resized partition /dev/sda9 Apr 24 23:33:26.756214 systemd[1]: Starting update-engine.service - Update Engine... Apr 24 23:33:26.801006 extend-filesystems[1619]: resize2fs 1.47.1 (20-May-2024) Apr 24 23:33:26.777485 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 24 23:33:26.790696 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 24 23:33:26.809838 jq[1622]: true Apr 24 23:33:26.807435 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 24 23:33:26.807682 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 24 23:33:26.807964 systemd[1]: motdgen.service: Deactivated successfully. Apr 24 23:33:26.810374 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 24 23:33:26.820721 update_engine[1614]: I20260424 23:33:26.820665 1614 main.cc:92] Flatcar Update Engine starting Apr 24 23:33:26.823450 update_engine[1614]: I20260424 23:33:26.823403 1614 update_check_scheduler.cc:74] Next update check in 6m12s Apr 24 23:33:26.823519 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 24 23:33:26.823773 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 24 23:33:26.862838 (ntainerd)[1631]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 24 23:33:26.875221 jq[1630]: true Apr 24 23:33:26.887510 tar[1628]: linux-amd64/LICENSE Apr 24 23:33:26.900302 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 31 scanned by (udev-worker) (1269) Apr 24 23:33:26.900346 tar[1628]: linux-amd64/helm Apr 24 23:33:26.908618 systemd[1]: Started update-engine.service - Update Engine. Apr 24 23:33:26.910961 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 24 23:33:26.910991 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 24 23:33:26.913378 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 24 23:33:26.913399 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 24 23:33:26.914181 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 24 23:33:26.925515 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 24 23:33:26.930344 systemd-networkd[1262]: eth1: Gained IPv6LL Apr 24 23:33:26.942984 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 24 23:33:26.960158 systemd[1]: Reached target network-online.target - Network is Online. Apr 24 23:33:26.974467 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 24 23:33:26.989302 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 24 23:33:26.994750 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Apr 24 23:33:27.003811 systemd-logind[1610]: New seat seat0. Apr 24 23:33:27.005338 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 24 23:33:27.011018 systemd-logind[1610]: Watching system buttons on /dev/input/event2 (Power Button) Apr 24 23:33:27.011040 systemd-logind[1610]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Apr 24 23:33:27.018950 systemd[1]: Started systemd-logind.service - User Login Management. Apr 24 23:33:27.074008 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 24 23:33:27.120715 systemd-networkd[1262]: eth0: Gained IPv6LL Apr 24 23:33:27.132701 locksmithd[1651]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 24 23:33:27.136464 sshd_keygen[1623]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 24 23:33:27.137917 bash[1678]: Updated "/home/core/.ssh/authorized_keys" Apr 24 23:33:27.139728 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 24 23:33:27.153298 systemd[1]: Starting sshkeys.service... Apr 24 23:33:27.169200 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Apr 24 23:33:27.179347 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Apr 24 23:33:27.203808 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 24 23:33:27.208180 containerd[1631]: time="2026-04-24T23:33:27.207443339Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Apr 24 23:33:27.210368 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 24 23:33:27.222330 coreos-metadata[1702]: Apr 24 23:33:27.222 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Apr 24 23:33:27.226851 coreos-metadata[1702]: Apr 24 23:33:27.226 INFO Fetch successful Apr 24 23:33:27.230217 unknown[1702]: wrote ssh authorized keys file for user: core Apr 24 23:33:27.234673 containerd[1631]: time="2026-04-24T23:33:27.231715320Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 24 23:33:27.234673 containerd[1631]: time="2026-04-24T23:33:27.233357341Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 24 23:33:27.234673 containerd[1631]: time="2026-04-24T23:33:27.233381671Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 24 23:33:27.234673 containerd[1631]: time="2026-04-24T23:33:27.233396071Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 24 23:33:27.234743 containerd[1631]: time="2026-04-24T23:33:27.234721392Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Apr 24 23:33:27.234764 containerd[1631]: time="2026-04-24T23:33:27.234742812Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Apr 24 23:33:27.235213 containerd[1631]: time="2026-04-24T23:33:27.234798372Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Apr 24 23:33:27.235213 containerd[1631]: time="2026-04-24T23:33:27.234811602Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 24 23:33:27.235213 containerd[1631]: time="2026-04-24T23:33:27.235034692Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 24 23:33:27.235213 containerd[1631]: time="2026-04-24T23:33:27.235046902Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 24 23:33:27.235213 containerd[1631]: time="2026-04-24T23:33:27.235055952Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Apr 24 23:33:27.235213 containerd[1631]: time="2026-04-24T23:33:27.235062622Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 24 23:33:27.235213 containerd[1631]: time="2026-04-24T23:33:27.235165522Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 24 23:33:27.235879 containerd[1631]: time="2026-04-24T23:33:27.235810333Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 24 23:33:27.236013 containerd[1631]: time="2026-04-24T23:33:27.235993133Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 24 23:33:27.236013 containerd[1631]: time="2026-04-24T23:33:27.236009373Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 24 23:33:27.237028 containerd[1631]: time="2026-04-24T23:33:27.236123163Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 24 23:33:27.237028 containerd[1631]: time="2026-04-24T23:33:27.236169013Z" level=info msg="metadata content store policy set" policy=shared Apr 24 23:33:27.246893 kernel: EXT4-fs (sda9): resized filesystem to 19393531 Apr 24 23:33:27.246600 systemd[1]: issuegen.service: Deactivated successfully. Apr 24 23:33:27.246864 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 24 23:33:27.258283 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 24 23:33:27.269718 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 24 23:33:27.276134 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 24 23:33:27.279826 extend-filesystems[1619]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Apr 24 23:33:27.279826 extend-filesystems[1619]: old_desc_blocks = 1, new_desc_blocks = 10 Apr 24 23:33:27.279826 extend-filesystems[1619]: The filesystem on /dev/sda9 is now 19393531 (4k) blocks long. Apr 24 23:33:27.288098 extend-filesystems[1591]: Resized filesystem in /dev/sda9 Apr 24 23:33:27.288098 extend-filesystems[1591]: Found sr0 Apr 24 23:33:27.288787 containerd[1631]: time="2026-04-24T23:33:27.285602554Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 24 23:33:27.288787 containerd[1631]: time="2026-04-24T23:33:27.285651084Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 24 23:33:27.288787 containerd[1631]: time="2026-04-24T23:33:27.285663654Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Apr 24 23:33:27.288787 containerd[1631]: time="2026-04-24T23:33:27.285677384Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Apr 24 23:33:27.288787 containerd[1631]: time="2026-04-24T23:33:27.285689085Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 24 23:33:27.288787 containerd[1631]: time="2026-04-24T23:33:27.286225095Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289741758Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289888578Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289903448Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289914488Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289924818Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289939468Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289948938Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289959508Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289970638Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289980208Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.289990398Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.290000238Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.290019628Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298206 containerd[1631]: time="2026-04-24T23:33:27.290030358Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.290560 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Apr 24 23:33:27.298491 update-ssh-keys[1718]: Updated "/home/core/.ssh/authorized_keys" Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.290040338Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.290050338Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.290061458Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291007629Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291029679Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291048679Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291059029Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291073969Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291892120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291911880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291922850Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291936950Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291953830Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291963460Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298668 containerd[1631]: time="2026-04-24T23:33:27.291971860Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 24 23:33:27.293375 systemd[1]: Reached target getty.target - Login Prompts. Apr 24 23:33:27.298916 containerd[1631]: time="2026-04-24T23:33:27.292013730Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 24 23:33:27.298916 containerd[1631]: time="2026-04-24T23:33:27.292026700Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Apr 24 23:33:27.298916 containerd[1631]: time="2026-04-24T23:33:27.292035150Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 24 23:33:27.298916 containerd[1631]: time="2026-04-24T23:33:27.292043130Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Apr 24 23:33:27.298916 containerd[1631]: time="2026-04-24T23:33:27.292049760Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.298916 containerd[1631]: time="2026-04-24T23:33:27.292058820Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Apr 24 23:33:27.298916 containerd[1631]: time="2026-04-24T23:33:27.292066640Z" level=info msg="NRI interface is disabled by configuration." Apr 24 23:33:27.298916 containerd[1631]: time="2026-04-24T23:33:27.292074680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.292315940Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.292360700Z" level=info msg="Connect containerd service" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.292389850Z" level=info msg="using legacy CRI server" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.292394980Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.292501280Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.292922121Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.296093043Z" level=info msg="Start subscribing containerd event" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.296344613Z" level=info msg="Start recovering state" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.296406483Z" level=info msg="Start event monitor" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.296424623Z" level=info msg="Start snapshots syncer" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.296431403Z" level=info msg="Start cni network conf syncer for default" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.296438863Z" level=info msg="Start streaming server" Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.297269824Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 24 23:33:27.299025 containerd[1631]: time="2026-04-24T23:33:27.297962635Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 24 23:33:27.299439 systemd[1]: Started containerd.service - containerd container runtime. Apr 24 23:33:27.300225 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 24 23:33:27.300490 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 24 23:33:27.304365 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Apr 24 23:33:27.311599 systemd[1]: Finished sshkeys.service. Apr 24 23:33:27.315169 containerd[1631]: time="2026-04-24T23:33:27.314461708Z" level=info msg="containerd successfully booted in 0.110257s" Apr 24 23:33:27.630182 tar[1628]: linux-amd64/README.md Apr 24 23:33:27.643575 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 24 23:33:28.024287 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 24 23:33:28.035509 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 24 23:33:28.035568 (kubelet)[1747]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 24 23:33:28.037877 systemd[1]: Startup finished in 7.416s (kernel) + 4.574s (userspace) = 11.990s. Apr 24 23:33:28.077802 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 24 23:33:28.084322 systemd[1]: Started sshd@0-95.217.179.164:22-4.175.71.9:49998.service - OpenSSH per-connection server daemon (4.175.71.9:49998). Apr 24 23:33:28.287758 sshd[1752]: Accepted publickey for core from 4.175.71.9 port 49998 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:33:28.291503 sshd[1752]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:33:28.304440 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 24 23:33:28.311409 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 24 23:33:28.316907 systemd-logind[1610]: New session 1 of user core. Apr 24 23:33:28.327587 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 24 23:33:28.336010 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 24 23:33:28.341547 (systemd)[1762]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 24 23:33:28.429966 systemd[1762]: Queued start job for default target default.target. Apr 24 23:33:28.430756 systemd[1762]: Created slice app.slice - User Application Slice. Apr 24 23:33:28.430775 systemd[1762]: Reached target paths.target - Paths. Apr 24 23:33:28.430786 systemd[1762]: Reached target timers.target - Timers. Apr 24 23:33:28.437209 systemd[1762]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 24 23:33:28.443363 systemd[1762]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 24 23:33:28.443415 systemd[1762]: Reached target sockets.target - Sockets. Apr 24 23:33:28.443427 systemd[1762]: Reached target basic.target - Basic System. Apr 24 23:33:28.443592 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 24 23:33:28.445129 systemd[1762]: Reached target default.target - Main User Target. Apr 24 23:33:28.445171 systemd[1762]: Startup finished in 97ms. Apr 24 23:33:28.450625 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 24 23:33:28.596602 kubelet[1747]: E0424 23:33:28.596491 1747 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 24 23:33:28.599464 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 24 23:33:28.599794 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 24 23:33:28.622513 systemd[1]: Started sshd@1-95.217.179.164:22-4.175.71.9:50004.service - OpenSSH per-connection server daemon (4.175.71.9:50004). Apr 24 23:33:28.823702 sshd[1777]: Accepted publickey for core from 4.175.71.9 port 50004 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:33:28.826706 sshd[1777]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:33:28.833938 systemd-logind[1610]: New session 2 of user core. Apr 24 23:33:28.836559 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 24 23:33:28.998368 sshd[1777]: pam_unix(sshd:session): session closed for user core Apr 24 23:33:29.004509 systemd[1]: sshd@1-95.217.179.164:22-4.175.71.9:50004.service: Deactivated successfully. Apr 24 23:33:29.011672 systemd-logind[1610]: Session 2 logged out. Waiting for processes to exit. Apr 24 23:33:29.013381 systemd[1]: session-2.scope: Deactivated successfully. Apr 24 23:33:29.015239 systemd-logind[1610]: Removed session 2. Apr 24 23:33:29.038519 systemd[1]: Started sshd@2-95.217.179.164:22-4.175.71.9:50020.service - OpenSSH per-connection server daemon (4.175.71.9:50020). Apr 24 23:33:29.249504 sshd[1785]: Accepted publickey for core from 4.175.71.9 port 50020 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:33:29.252171 sshd[1785]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:33:29.260144 systemd-logind[1610]: New session 3 of user core. Apr 24 23:33:29.265621 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 24 23:33:29.412990 sshd[1785]: pam_unix(sshd:session): session closed for user core Apr 24 23:33:29.420769 systemd[1]: sshd@2-95.217.179.164:22-4.175.71.9:50020.service: Deactivated successfully. Apr 24 23:33:29.425712 systemd[1]: session-3.scope: Deactivated successfully. Apr 24 23:33:29.426910 systemd-logind[1610]: Session 3 logged out. Waiting for processes to exit. Apr 24 23:33:29.428854 systemd-logind[1610]: Removed session 3. Apr 24 23:33:29.453442 systemd[1]: Started sshd@3-95.217.179.164:22-4.175.71.9:50030.service - OpenSSH per-connection server daemon (4.175.71.9:50030). Apr 24 23:33:29.667213 sshd[1793]: Accepted publickey for core from 4.175.71.9 port 50030 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:33:29.668809 sshd[1793]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:33:29.676738 systemd-logind[1610]: New session 4 of user core. Apr 24 23:33:29.686603 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 24 23:33:29.840348 sshd[1793]: pam_unix(sshd:session): session closed for user core Apr 24 23:33:29.847615 systemd[1]: sshd@3-95.217.179.164:22-4.175.71.9:50030.service: Deactivated successfully. Apr 24 23:33:29.853013 systemd[1]: session-4.scope: Deactivated successfully. Apr 24 23:33:29.854746 systemd-logind[1610]: Session 4 logged out. Waiting for processes to exit. Apr 24 23:33:29.856434 systemd-logind[1610]: Removed session 4. Apr 24 23:33:29.881450 systemd[1]: Started sshd@4-95.217.179.164:22-4.175.71.9:50036.service - OpenSSH per-connection server daemon (4.175.71.9:50036). Apr 24 23:33:30.103447 sshd[1801]: Accepted publickey for core from 4.175.71.9 port 50036 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:33:30.106807 sshd[1801]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:33:30.114568 systemd-logind[1610]: New session 5 of user core. Apr 24 23:33:30.121548 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 24 23:33:30.258449 sudo[1805]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 24 23:33:30.259724 sudo[1805]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 24 23:33:30.291004 sudo[1805]: pam_unix(sudo:session): session closed for user root Apr 24 23:33:30.323059 sshd[1801]: pam_unix(sshd:session): session closed for user core Apr 24 23:33:30.329061 systemd[1]: sshd@4-95.217.179.164:22-4.175.71.9:50036.service: Deactivated successfully. Apr 24 23:33:30.336367 systemd-logind[1610]: Session 5 logged out. Waiting for processes to exit. Apr 24 23:33:30.337614 systemd[1]: session-5.scope: Deactivated successfully. Apr 24 23:33:30.339517 systemd-logind[1610]: Removed session 5. Apr 24 23:33:30.365734 systemd[1]: Started sshd@5-95.217.179.164:22-4.175.71.9:50046.service - OpenSSH per-connection server daemon (4.175.71.9:50046). Apr 24 23:33:30.573878 sshd[1810]: Accepted publickey for core from 4.175.71.9 port 50046 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:33:30.577362 sshd[1810]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:33:30.584873 systemd-logind[1610]: New session 6 of user core. Apr 24 23:33:30.590571 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 24 23:33:30.716945 sudo[1815]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 24 23:33:30.717975 sudo[1815]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 24 23:33:30.724510 sudo[1815]: pam_unix(sudo:session): session closed for user root Apr 24 23:33:30.736124 sudo[1814]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Apr 24 23:33:30.736825 sudo[1814]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 24 23:33:30.757475 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Apr 24 23:33:30.772620 auditctl[1818]: No rules Apr 24 23:33:30.773760 systemd[1]: audit-rules.service: Deactivated successfully. Apr 24 23:33:30.774262 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Apr 24 23:33:30.788597 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 24 23:33:30.836500 augenrules[1837]: No rules Apr 24 23:33:30.840504 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 24 23:33:30.845203 sudo[1814]: pam_unix(sudo:session): session closed for user root Apr 24 23:33:30.877648 sshd[1810]: pam_unix(sshd:session): session closed for user core Apr 24 23:33:30.883662 systemd-logind[1610]: Session 6 logged out. Waiting for processes to exit. Apr 24 23:33:30.885173 systemd[1]: sshd@5-95.217.179.164:22-4.175.71.9:50046.service: Deactivated successfully. Apr 24 23:33:30.889859 systemd[1]: session-6.scope: Deactivated successfully. Apr 24 23:33:30.891832 systemd-logind[1610]: Removed session 6. Apr 24 23:33:30.920469 systemd[1]: Started sshd@6-95.217.179.164:22-4.175.71.9:50052.service - OpenSSH per-connection server daemon (4.175.71.9:50052). Apr 24 23:33:31.140573 sshd[1846]: Accepted publickey for core from 4.175.71.9 port 50052 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:33:31.142881 sshd[1846]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:33:31.151192 systemd-logind[1610]: New session 7 of user core. Apr 24 23:33:31.159270 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 24 23:33:31.284531 sudo[1850]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 24 23:33:31.285348 sudo[1850]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 24 23:33:31.566268 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 24 23:33:31.569402 (dockerd)[1866]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 24 23:33:31.802640 dockerd[1866]: time="2026-04-24T23:33:31.802571417Z" level=info msg="Starting up" Apr 24 23:33:31.944750 dockerd[1866]: time="2026-04-24T23:33:31.944261825Z" level=info msg="Loading containers: start." Apr 24 23:33:32.050259 kernel: Initializing XFRM netlink socket Apr 24 23:33:32.121852 systemd-networkd[1262]: docker0: Link UP Apr 24 23:33:32.134277 dockerd[1866]: time="2026-04-24T23:33:32.134232214Z" level=info msg="Loading containers: done." Apr 24 23:33:32.149161 dockerd[1866]: time="2026-04-24T23:33:32.149109876Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 24 23:33:32.149299 dockerd[1866]: time="2026-04-24T23:33:32.149220496Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Apr 24 23:33:32.149371 dockerd[1866]: time="2026-04-24T23:33:32.149349426Z" level=info msg="Daemon has completed initialization" Apr 24 23:33:32.175767 dockerd[1866]: time="2026-04-24T23:33:32.175713088Z" level=info msg="API listen on /run/docker.sock" Apr 24 23:33:32.176049 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 24 23:33:32.625708 containerd[1631]: time="2026-04-24T23:33:32.625672273Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.11\"" Apr 24 23:33:33.184152 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1755185988.mount: Deactivated successfully. Apr 24 23:33:34.381776 containerd[1631]: time="2026-04-24T23:33:34.381734046Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:34.382484 containerd[1631]: time="2026-04-24T23:33:34.382462867Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.11: active requests=0, bytes read=30194089" Apr 24 23:33:34.383308 containerd[1631]: time="2026-04-24T23:33:34.383286527Z" level=info msg="ImageCreate event name:\"sha256:7ea99c30f23b106a042b6c46e565fddb42b20bbe58ba6852e562eed03477aec2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:34.386639 containerd[1631]: time="2026-04-24T23:33:34.386615860Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:18e9f2b6e4d67c24941e14b2d41ec0aa6e5f628e39f2ef2163e176de85bbe39e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:34.387472 containerd[1631]: time="2026-04-24T23:33:34.387447721Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.11\" with image id \"sha256:7ea99c30f23b106a042b6c46e565fddb42b20bbe58ba6852e562eed03477aec2\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:18e9f2b6e4d67c24941e14b2d41ec0aa6e5f628e39f2ef2163e176de85bbe39e\", size \"30190588\" in 1.761744358s" Apr 24 23:33:34.387515 containerd[1631]: time="2026-04-24T23:33:34.387477171Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.11\" returns image reference \"sha256:7ea99c30f23b106a042b6c46e565fddb42b20bbe58ba6852e562eed03477aec2\"" Apr 24 23:33:34.388174 containerd[1631]: time="2026-04-24T23:33:34.388154051Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.11\"" Apr 24 23:33:35.535757 containerd[1631]: time="2026-04-24T23:33:35.535716607Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:35.536877 containerd[1631]: time="2026-04-24T23:33:35.536847618Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.11: active requests=0, bytes read=26171469" Apr 24 23:33:35.538092 containerd[1631]: time="2026-04-24T23:33:35.538052599Z" level=info msg="ImageCreate event name:\"sha256:c75dc8a6c47e2f7491fa2e367879f53c6f46053066e6b7135df4b154ddd94a1f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:35.540151 containerd[1631]: time="2026-04-24T23:33:35.540123821Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:7579451c5b3c2715da4a263c5d80a3367a24fdc12e86fde6851674d567d1dfb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:35.541278 containerd[1631]: time="2026-04-24T23:33:35.540802142Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.11\" with image id \"sha256:c75dc8a6c47e2f7491fa2e367879f53c6f46053066e6b7135df4b154ddd94a1f\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:7579451c5b3c2715da4a263c5d80a3367a24fdc12e86fde6851674d567d1dfb2\", size \"27737794\" in 1.152627391s" Apr 24 23:33:35.541278 containerd[1631]: time="2026-04-24T23:33:35.540825342Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.11\" returns image reference \"sha256:c75dc8a6c47e2f7491fa2e367879f53c6f46053066e6b7135df4b154ddd94a1f\"" Apr 24 23:33:35.541278 containerd[1631]: time="2026-04-24T23:33:35.541153172Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.11\"" Apr 24 23:33:36.579670 containerd[1631]: time="2026-04-24T23:33:36.579623997Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:36.580884 containerd[1631]: time="2026-04-24T23:33:36.580761328Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.11: active requests=0, bytes read=20289778" Apr 24 23:33:36.582544 containerd[1631]: time="2026-04-24T23:33:36.581580629Z" level=info msg="ImageCreate event name:\"sha256:3febad3451e2d599688a8ad13d19d03c48c9054be209342c748fac2bb6c56f97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:36.583696 containerd[1631]: time="2026-04-24T23:33:36.583667460Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:5506f0f94c4d9aeb071664893aabc12166bcb7f775008a6fff02d004e6091d28\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:36.584602 containerd[1631]: time="2026-04-24T23:33:36.584466301Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.11\" with image id \"sha256:3febad3451e2d599688a8ad13d19d03c48c9054be209342c748fac2bb6c56f97\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:5506f0f94c4d9aeb071664893aabc12166bcb7f775008a6fff02d004e6091d28\", size \"21856121\" in 1.043298349s" Apr 24 23:33:36.584602 containerd[1631]: time="2026-04-24T23:33:36.584488931Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.11\" returns image reference \"sha256:3febad3451e2d599688a8ad13d19d03c48c9054be209342c748fac2bb6c56f97\"" Apr 24 23:33:36.585139 containerd[1631]: time="2026-04-24T23:33:36.585112322Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.11\"" Apr 24 23:33:37.555133 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3988339418.mount: Deactivated successfully. Apr 24 23:33:37.876267 containerd[1631]: time="2026-04-24T23:33:37.876170057Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:37.877400 containerd[1631]: time="2026-04-24T23:33:37.877314238Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.11: active requests=0, bytes read=32010739" Apr 24 23:33:37.878611 containerd[1631]: time="2026-04-24T23:33:37.878274279Z" level=info msg="ImageCreate event name:\"sha256:4ce1332df15d2a0b1c2d3b18292afb4ff670070401211daebb00b7293b26f6d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:37.879893 containerd[1631]: time="2026-04-24T23:33:37.879865290Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8d18637b5c5f58a4ca0163d3cf184e53d4c522963c242860562be7cb25e9303e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:37.880397 containerd[1631]: time="2026-04-24T23:33:37.880277171Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.11\" with image id \"sha256:4ce1332df15d2a0b1c2d3b18292afb4ff670070401211daebb00b7293b26f6d0\", repo tag \"registry.k8s.io/kube-proxy:v1.33.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:8d18637b5c5f58a4ca0163d3cf184e53d4c522963c242860562be7cb25e9303e\", size \"32009730\" in 1.295137849s" Apr 24 23:33:37.880397 containerd[1631]: time="2026-04-24T23:33:37.880301521Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.11\" returns image reference \"sha256:4ce1332df15d2a0b1c2d3b18292afb4ff670070401211daebb00b7293b26f6d0\"" Apr 24 23:33:37.880646 containerd[1631]: time="2026-04-24T23:33:37.880629251Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Apr 24 23:33:38.404708 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2710763657.mount: Deactivated successfully. Apr 24 23:33:38.794638 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 24 23:33:38.806345 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 24 23:33:38.963763 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 24 23:33:38.967489 (kubelet)[2143]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 24 23:33:39.003964 kubelet[2143]: E0424 23:33:39.003926 2143 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 24 23:33:39.008725 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 24 23:33:39.009932 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 24 23:33:39.211903 containerd[1631]: time="2026-04-24T23:33:39.211798390Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:39.213271 containerd[1631]: time="2026-04-24T23:33:39.213100311Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942332" Apr 24 23:33:39.215407 containerd[1631]: time="2026-04-24T23:33:39.214119052Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:39.216587 containerd[1631]: time="2026-04-24T23:33:39.216375574Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:39.217050 containerd[1631]: time="2026-04-24T23:33:39.217030144Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.336379943s" Apr 24 23:33:39.217097 containerd[1631]: time="2026-04-24T23:33:39.217054244Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Apr 24 23:33:39.217835 containerd[1631]: time="2026-04-24T23:33:39.217814795Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Apr 24 23:33:39.716341 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2599380581.mount: Deactivated successfully. Apr 24 23:33:39.719637 containerd[1631]: time="2026-04-24T23:33:39.719559793Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:39.720524 containerd[1631]: time="2026-04-24T23:33:39.720497824Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321160" Apr 24 23:33:39.721248 containerd[1631]: time="2026-04-24T23:33:39.721203464Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:39.723602 containerd[1631]: time="2026-04-24T23:33:39.723554096Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:39.726137 containerd[1631]: time="2026-04-24T23:33:39.726071828Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 508.230913ms" Apr 24 23:33:39.726253 containerd[1631]: time="2026-04-24T23:33:39.726144188Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Apr 24 23:33:39.726967 containerd[1631]: time="2026-04-24T23:33:39.726908719Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Apr 24 23:33:40.285673 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2096878609.mount: Deactivated successfully. Apr 24 23:33:41.176548 containerd[1631]: time="2026-04-24T23:33:41.176500896Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:41.177618 containerd[1631]: time="2026-04-24T23:33:41.177579167Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=23719532" Apr 24 23:33:41.178642 containerd[1631]: time="2026-04-24T23:33:41.178627138Z" level=info msg="ImageCreate event name:\"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:41.181162 containerd[1631]: time="2026-04-24T23:33:41.181129650Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:41.183102 containerd[1631]: time="2026-04-24T23:33:41.181851811Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"23716032\" in 1.454906492s" Apr 24 23:33:41.183102 containerd[1631]: time="2026-04-24T23:33:41.181872671Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\"" Apr 24 23:33:43.113481 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 24 23:33:43.124239 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 24 23:33:43.145991 systemd[1]: Reloading requested from client PID 2248 ('systemctl') (unit session-7.scope)... Apr 24 23:33:43.146001 systemd[1]: Reloading... Apr 24 23:33:43.245123 zram_generator::config[2288]: No configuration found. Apr 24 23:33:43.342921 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 24 23:33:43.403862 systemd[1]: Reloading finished in 257 ms. Apr 24 23:33:43.458241 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Apr 24 23:33:43.458332 systemd[1]: kubelet.service: Failed with result 'signal'. Apr 24 23:33:43.458722 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 24 23:33:43.466580 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 24 23:33:43.591217 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 24 23:33:43.603831 (kubelet)[2353]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 24 23:33:43.644112 kubelet[2353]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 24 23:33:43.644112 kubelet[2353]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 24 23:33:43.644112 kubelet[2353]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 24 23:33:43.644112 kubelet[2353]: I0424 23:33:43.643501 2353 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 24 23:33:44.357006 kubelet[2353]: I0424 23:33:44.356947 2353 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 24 23:33:44.357006 kubelet[2353]: I0424 23:33:44.356969 2353 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 24 23:33:44.357289 kubelet[2353]: I0424 23:33:44.357146 2353 server.go:956] "Client rotation is on, will bootstrap in background" Apr 24 23:33:44.383189 kubelet[2353]: I0424 23:33:44.382571 2353 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 24 23:33:44.385431 kubelet[2353]: E0424 23:33:44.385372 2353 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://95.217.179.164:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 95.217.179.164:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 24 23:33:44.391466 kubelet[2353]: E0424 23:33:44.391434 2353 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 24 23:33:44.391466 kubelet[2353]: I0424 23:33:44.391453 2353 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 24 23:33:44.394530 kubelet[2353]: I0424 23:33:44.394488 2353 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 24 23:33:44.394897 kubelet[2353]: I0424 23:33:44.394855 2353 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 24 23:33:44.395002 kubelet[2353]: I0424 23:33:44.394877 2353 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-6-n-e88c017743","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Apr 24 23:33:44.395002 kubelet[2353]: I0424 23:33:44.394988 2353 topology_manager.go:138] "Creating topology manager with none policy" Apr 24 23:33:44.395002 kubelet[2353]: I0424 23:33:44.394994 2353 container_manager_linux.go:303] "Creating device plugin manager" Apr 24 23:33:44.395216 kubelet[2353]: I0424 23:33:44.395102 2353 state_mem.go:36] "Initialized new in-memory state store" Apr 24 23:33:44.398644 kubelet[2353]: I0424 23:33:44.398607 2353 kubelet.go:480] "Attempting to sync node with API server" Apr 24 23:33:44.398644 kubelet[2353]: I0424 23:33:44.398622 2353 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 24 23:33:44.398644 kubelet[2353]: I0424 23:33:44.398643 2353 kubelet.go:386] "Adding apiserver pod source" Apr 24 23:33:44.398939 kubelet[2353]: I0424 23:33:44.398662 2353 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 24 23:33:44.404135 kubelet[2353]: E0424 23:33:44.403804 2353 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://95.217.179.164:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-6-n-e88c017743&limit=500&resourceVersion=0\": dial tcp 95.217.179.164:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 24 23:33:44.404135 kubelet[2353]: I0424 23:33:44.403876 2353 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 24 23:33:44.404261 kubelet[2353]: I0424 23:33:44.404245 2353 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 24 23:33:44.404857 kubelet[2353]: W0424 23:33:44.404837 2353 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 24 23:33:44.405222 kubelet[2353]: E0424 23:33:44.405199 2353 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://95.217.179.164:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 95.217.179.164:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 24 23:33:44.408759 kubelet[2353]: I0424 23:33:44.408737 2353 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 24 23:33:44.408842 kubelet[2353]: I0424 23:33:44.408766 2353 server.go:1289] "Started kubelet" Apr 24 23:33:44.409845 kubelet[2353]: I0424 23:33:44.409812 2353 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 24 23:33:44.418833 kubelet[2353]: I0424 23:33:44.417462 2353 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 24 23:33:44.418833 kubelet[2353]: E0424 23:33:44.417596 2353 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-e88c017743\" not found" Apr 24 23:33:44.418833 kubelet[2353]: I0424 23:33:44.418006 2353 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 24 23:33:44.419393 kubelet[2353]: I0424 23:33:44.419374 2353 server.go:317] "Adding debug handlers to kubelet server" Apr 24 23:33:44.420062 kubelet[2353]: I0424 23:33:44.420035 2353 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 24 23:33:44.420148 kubelet[2353]: I0424 23:33:44.420071 2353 reconciler.go:26] "Reconciler: start to sync state" Apr 24 23:33:44.423484 kubelet[2353]: I0424 23:33:44.423118 2353 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 24 23:33:44.423484 kubelet[2353]: I0424 23:33:44.423293 2353 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 24 23:33:44.427290 kubelet[2353]: I0424 23:33:44.426505 2353 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 24 23:33:44.427290 kubelet[2353]: I0424 23:33:44.426742 2353 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 24 23:33:44.427290 kubelet[2353]: E0424 23:33:44.427000 2353 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://95.217.179.164:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-e88c017743?timeout=10s\": dial tcp 95.217.179.164:6443: connect: connection refused" interval="200ms" Apr 24 23:33:44.428851 kubelet[2353]: I0424 23:33:44.428055 2353 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 24 23:33:44.428851 kubelet[2353]: I0424 23:33:44.428076 2353 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 24 23:33:44.428851 kubelet[2353]: I0424 23:33:44.428110 2353 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 24 23:33:44.428851 kubelet[2353]: I0424 23:33:44.428117 2353 kubelet.go:2436] "Starting kubelet main sync loop" Apr 24 23:33:44.428851 kubelet[2353]: E0424 23:33:44.428146 2353 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 24 23:33:44.429657 kubelet[2353]: E0424 23:33:44.427197 2353 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://95.217.179.164:6443/api/v1/namespaces/default/events\": dial tcp 95.217.179.164:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081-3-6-n-e88c017743.18a96f0bca997877 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-3-6-n-e88c017743,UID:ci-4081-3-6-n-e88c017743,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-3-6-n-e88c017743,},FirstTimestamp:2026-04-24 23:33:44.408750199 +0000 UTC m=+0.799054947,LastTimestamp:2026-04-24 23:33:44.408750199 +0000 UTC m=+0.799054947,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-6-n-e88c017743,}" Apr 24 23:33:44.429819 kubelet[2353]: E0424 23:33:44.429803 2353 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://95.217.179.164:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 95.217.179.164:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 24 23:33:44.431667 kubelet[2353]: I0424 23:33:44.431641 2353 factory.go:223] Registration of the systemd container factory successfully Apr 24 23:33:44.431793 kubelet[2353]: I0424 23:33:44.431773 2353 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 24 23:33:44.432905 kubelet[2353]: E0424 23:33:44.432850 2353 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 24 23:33:44.433913 kubelet[2353]: I0424 23:33:44.433079 2353 factory.go:223] Registration of the containerd container factory successfully Apr 24 23:33:44.436252 kubelet[2353]: E0424 23:33:44.436237 2353 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://95.217.179.164:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 95.217.179.164:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 24 23:33:44.454131 kubelet[2353]: I0424 23:33:44.453937 2353 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 24 23:33:44.454131 kubelet[2353]: I0424 23:33:44.453946 2353 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 24 23:33:44.454131 kubelet[2353]: I0424 23:33:44.453959 2353 state_mem.go:36] "Initialized new in-memory state store" Apr 24 23:33:44.455989 kubelet[2353]: I0424 23:33:44.455966 2353 policy_none.go:49] "None policy: Start" Apr 24 23:33:44.455989 kubelet[2353]: I0424 23:33:44.455981 2353 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 24 23:33:44.455989 kubelet[2353]: I0424 23:33:44.455991 2353 state_mem.go:35] "Initializing new in-memory state store" Apr 24 23:33:44.459922 kubelet[2353]: E0424 23:33:44.459901 2353 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 24 23:33:44.460047 kubelet[2353]: I0424 23:33:44.460033 2353 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 24 23:33:44.460098 kubelet[2353]: I0424 23:33:44.460044 2353 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 24 23:33:44.461423 kubelet[2353]: I0424 23:33:44.461363 2353 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 24 23:33:44.464489 kubelet[2353]: E0424 23:33:44.464468 2353 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 24 23:33:44.464542 kubelet[2353]: E0424 23:33:44.464497 2353 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081-3-6-n-e88c017743\" not found" Apr 24 23:33:44.541562 kubelet[2353]: E0424 23:33:44.540943 2353 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-e88c017743\" not found" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.547300 kubelet[2353]: E0424 23:33:44.546991 2353 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-e88c017743\" not found" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.556141 kubelet[2353]: E0424 23:33:44.554429 2353 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-e88c017743\" not found" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.562700 kubelet[2353]: I0424 23:33:44.562632 2353 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.563037 kubelet[2353]: E0424 23:33:44.562999 2353 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://95.217.179.164:6443/api/v1/nodes\": dial tcp 95.217.179.164:6443: connect: connection refused" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.621767 kubelet[2353]: I0424 23:33:44.621580 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9e69cda47ba3742cdfd1088ebeda4528-k8s-certs\") pod \"kube-apiserver-ci-4081-3-6-n-e88c017743\" (UID: \"9e69cda47ba3742cdfd1088ebeda4528\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.621767 kubelet[2353]: I0424 23:33:44.621635 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.621767 kubelet[2353]: I0424 23:33:44.621667 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.621767 kubelet[2353]: I0424 23:33:44.621703 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9e69cda47ba3742cdfd1088ebeda4528-ca-certs\") pod \"kube-apiserver-ci-4081-3-6-n-e88c017743\" (UID: \"9e69cda47ba3742cdfd1088ebeda4528\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.622013 kubelet[2353]: I0424 23:33:44.621773 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9e69cda47ba3742cdfd1088ebeda4528-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-6-n-e88c017743\" (UID: \"9e69cda47ba3742cdfd1088ebeda4528\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.622013 kubelet[2353]: I0424 23:33:44.621831 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-ca-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.622013 kubelet[2353]: I0424 23:33:44.621860 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.622013 kubelet[2353]: I0424 23:33:44.621884 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.622013 kubelet[2353]: I0424 23:33:44.621909 2353 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0fb18c868bc1dc81e6398608291507ac-kubeconfig\") pod \"kube-scheduler-ci-4081-3-6-n-e88c017743\" (UID: \"0fb18c868bc1dc81e6398608291507ac\") " pod="kube-system/kube-scheduler-ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.628212 kubelet[2353]: E0424 23:33:44.628166 2353 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://95.217.179.164:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-e88c017743?timeout=10s\": dial tcp 95.217.179.164:6443: connect: connection refused" interval="400ms" Apr 24 23:33:44.766191 kubelet[2353]: I0424 23:33:44.766149 2353 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.767103 kubelet[2353]: E0424 23:33:44.767025 2353 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://95.217.179.164:6443/api/v1/nodes\": dial tcp 95.217.179.164:6443: connect: connection refused" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:44.843055 containerd[1631]: time="2026-04-24T23:33:44.842902731Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-6-n-e88c017743,Uid:9e69cda47ba3742cdfd1088ebeda4528,Namespace:kube-system,Attempt:0,}" Apr 24 23:33:44.848563 containerd[1631]: time="2026-04-24T23:33:44.848514876Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-6-n-e88c017743,Uid:5fd927beb9b34649f92f550ce577c85b,Namespace:kube-system,Attempt:0,}" Apr 24 23:33:44.859941 containerd[1631]: time="2026-04-24T23:33:44.859574395Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-6-n-e88c017743,Uid:0fb18c868bc1dc81e6398608291507ac,Namespace:kube-system,Attempt:0,}" Apr 24 23:33:45.030144 kubelet[2353]: E0424 23:33:45.029931 2353 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://95.217.179.164:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-e88c017743?timeout=10s\": dial tcp 95.217.179.164:6443: connect: connection refused" interval="800ms" Apr 24 23:33:45.169893 kubelet[2353]: I0424 23:33:45.169552 2353 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:45.170051 kubelet[2353]: E0424 23:33:45.169962 2353 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://95.217.179.164:6443/api/v1/nodes\": dial tcp 95.217.179.164:6443: connect: connection refused" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:45.331694 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount358080254.mount: Deactivated successfully. Apr 24 23:33:45.339140 containerd[1631]: time="2026-04-24T23:33:45.337745783Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 24 23:33:45.339401 containerd[1631]: time="2026-04-24T23:33:45.339364484Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 24 23:33:45.340999 containerd[1631]: time="2026-04-24T23:33:45.340903376Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 24 23:33:45.341573 containerd[1631]: time="2026-04-24T23:33:45.341500076Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312078" Apr 24 23:33:45.345129 containerd[1631]: time="2026-04-24T23:33:45.344117898Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 24 23:33:45.345989 containerd[1631]: time="2026-04-24T23:33:45.345952980Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 24 23:33:45.348016 containerd[1631]: time="2026-04-24T23:33:45.347953562Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 24 23:33:45.349669 containerd[1631]: time="2026-04-24T23:33:45.349616893Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 24 23:33:45.354179 containerd[1631]: time="2026-04-24T23:33:45.354136957Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 505.504971ms" Apr 24 23:33:45.358339 containerd[1631]: time="2026-04-24T23:33:45.358274330Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 515.293179ms" Apr 24 23:33:45.360162 containerd[1631]: time="2026-04-24T23:33:45.359920052Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 500.217537ms" Apr 24 23:33:45.416729 kubelet[2353]: E0424 23:33:45.416687 2353 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://95.217.179.164:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 95.217.179.164:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 24 23:33:45.439294 containerd[1631]: time="2026-04-24T23:33:45.439211418Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:33:45.439294 containerd[1631]: time="2026-04-24T23:33:45.439251318Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:33:45.439294 containerd[1631]: time="2026-04-24T23:33:45.439273768Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:45.439660 containerd[1631]: time="2026-04-24T23:33:45.439612278Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:45.447316 containerd[1631]: time="2026-04-24T23:33:45.447255244Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:33:45.447316 containerd[1631]: time="2026-04-24T23:33:45.447287824Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:33:45.447421 containerd[1631]: time="2026-04-24T23:33:45.447306674Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:45.447512 containerd[1631]: time="2026-04-24T23:33:45.447485875Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:45.448465 containerd[1631]: time="2026-04-24T23:33:45.448047605Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:33:45.452443 containerd[1631]: time="2026-04-24T23:33:45.452397399Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:33:45.452555 containerd[1631]: time="2026-04-24T23:33:45.452431809Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:45.452653 containerd[1631]: time="2026-04-24T23:33:45.452625569Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:45.512099 containerd[1631]: time="2026-04-24T23:33:45.512048638Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-6-n-e88c017743,Uid:9e69cda47ba3742cdfd1088ebeda4528,Namespace:kube-system,Attempt:0,} returns sandbox id \"7145f8592ee47d83c89612ace6deaafd26daa74c9d27dca466ae2ef08fa258ba\"" Apr 24 23:33:45.517144 containerd[1631]: time="2026-04-24T23:33:45.517121523Z" level=info msg="CreateContainer within sandbox \"7145f8592ee47d83c89612ace6deaafd26daa74c9d27dca466ae2ef08fa258ba\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 24 23:33:45.518298 containerd[1631]: time="2026-04-24T23:33:45.518284063Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-6-n-e88c017743,Uid:5fd927beb9b34649f92f550ce577c85b,Namespace:kube-system,Attempt:0,} returns sandbox id \"f54fb2fca546265cbee3dcbe31874be2bdba59cf461901e10a9c1711863b52fb\"" Apr 24 23:33:45.522797 containerd[1631]: time="2026-04-24T23:33:45.522688687Z" level=info msg="CreateContainer within sandbox \"f54fb2fca546265cbee3dcbe31874be2bdba59cf461901e10a9c1711863b52fb\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 24 23:33:45.525116 containerd[1631]: time="2026-04-24T23:33:45.525063619Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-6-n-e88c017743,Uid:0fb18c868bc1dc81e6398608291507ac,Namespace:kube-system,Attempt:0,} returns sandbox id \"788cfc01e78ddecdfecdaf0a48c7d7cfa593f71c8927e1b3f635ed9199d38df9\"" Apr 24 23:33:45.528052 containerd[1631]: time="2026-04-24T23:33:45.528007562Z" level=info msg="CreateContainer within sandbox \"788cfc01e78ddecdfecdaf0a48c7d7cfa593f71c8927e1b3f635ed9199d38df9\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 24 23:33:45.534698 containerd[1631]: time="2026-04-24T23:33:45.534674677Z" level=info msg="CreateContainer within sandbox \"7145f8592ee47d83c89612ace6deaafd26daa74c9d27dca466ae2ef08fa258ba\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"49c6d76be1d573a2f7babb6ca50e7cda777c8e56e96f7304ad270d79760867fe\"" Apr 24 23:33:45.535350 containerd[1631]: time="2026-04-24T23:33:45.535314288Z" level=info msg="StartContainer for \"49c6d76be1d573a2f7babb6ca50e7cda777c8e56e96f7304ad270d79760867fe\"" Apr 24 23:33:45.538560 containerd[1631]: time="2026-04-24T23:33:45.538477070Z" level=info msg="CreateContainer within sandbox \"f54fb2fca546265cbee3dcbe31874be2bdba59cf461901e10a9c1711863b52fb\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"94229b8631e59bfe3a1dc54f88abcb1f153e15884a228cf13cfb1d456c99f77d\"" Apr 24 23:33:45.538908 containerd[1631]: time="2026-04-24T23:33:45.538847821Z" level=info msg="StartContainer for \"94229b8631e59bfe3a1dc54f88abcb1f153e15884a228cf13cfb1d456c99f77d\"" Apr 24 23:33:45.541848 containerd[1631]: time="2026-04-24T23:33:45.541830533Z" level=info msg="CreateContainer within sandbox \"788cfc01e78ddecdfecdaf0a48c7d7cfa593f71c8927e1b3f635ed9199d38df9\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c59e36393a4c3113197c00ce89282d2b2c6fedd28206f840df82a8ae54fdd6a7\"" Apr 24 23:33:45.542254 containerd[1631]: time="2026-04-24T23:33:45.542208833Z" level=info msg="StartContainer for \"c59e36393a4c3113197c00ce89282d2b2c6fedd28206f840df82a8ae54fdd6a7\"" Apr 24 23:33:45.634806 containerd[1631]: time="2026-04-24T23:33:45.634443410Z" level=info msg="StartContainer for \"49c6d76be1d573a2f7babb6ca50e7cda777c8e56e96f7304ad270d79760867fe\" returns successfully" Apr 24 23:33:45.635979 containerd[1631]: time="2026-04-24T23:33:45.635077811Z" level=info msg="StartContainer for \"c59e36393a4c3113197c00ce89282d2b2c6fedd28206f840df82a8ae54fdd6a7\" returns successfully" Apr 24 23:33:45.655640 containerd[1631]: time="2026-04-24T23:33:45.655601928Z" level=info msg="StartContainer for \"94229b8631e59bfe3a1dc54f88abcb1f153e15884a228cf13cfb1d456c99f77d\" returns successfully" Apr 24 23:33:45.973183 kubelet[2353]: I0424 23:33:45.972930 2353 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.459557 kubelet[2353]: E0424 23:33:46.459401 2353 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-e88c017743\" not found" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.465103 kubelet[2353]: E0424 23:33:46.464448 2353 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-e88c017743\" not found" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.469308 kubelet[2353]: E0424 23:33:46.469294 2353 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-e88c017743\" not found" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.507630 kubelet[2353]: E0424 23:33:46.507603 2353 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081-3-6-n-e88c017743\" not found" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.567653 kubelet[2353]: I0424 23:33:46.567436 2353 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.567653 kubelet[2353]: E0424 23:33:46.567466 2353 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-4081-3-6-n-e88c017743\": node \"ci-4081-3-6-n-e88c017743\" not found" Apr 24 23:33:46.618493 kubelet[2353]: I0424 23:33:46.618175 2353 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.678510 kubelet[2353]: E0424 23:33:46.678465 2353 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-e88c017743\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.678510 kubelet[2353]: I0424 23:33:46.678487 2353 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.681909 kubelet[2353]: E0424 23:33:46.681860 2353 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.681909 kubelet[2353]: I0424 23:33:46.681893 2353 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-e88c017743" Apr 24 23:33:46.683161 kubelet[2353]: E0424 23:33:46.683125 2353 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-6-n-e88c017743\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081-3-6-n-e88c017743" Apr 24 23:33:47.404254 kubelet[2353]: I0424 23:33:47.402458 2353 apiserver.go:52] "Watching apiserver" Apr 24 23:33:47.421015 kubelet[2353]: I0424 23:33:47.420959 2353 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 24 23:33:47.472800 kubelet[2353]: I0424 23:33:47.470316 2353 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:47.473521 kubelet[2353]: I0424 23:33:47.473397 2353 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-e88c017743" Apr 24 23:33:48.552184 systemd[1]: Reloading requested from client PID 2637 ('systemctl') (unit session-7.scope)... Apr 24 23:33:48.552212 systemd[1]: Reloading... Apr 24 23:33:48.632144 zram_generator::config[2677]: No configuration found. Apr 24 23:33:48.749969 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 24 23:33:48.817696 systemd[1]: Reloading finished in 264 ms. Apr 24 23:33:48.849635 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 24 23:33:48.874492 systemd[1]: kubelet.service: Deactivated successfully. Apr 24 23:33:48.874908 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 24 23:33:48.880523 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 24 23:33:49.012212 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 24 23:33:49.021558 (kubelet)[2738]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 24 23:33:49.058276 kubelet[2738]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 24 23:33:49.058276 kubelet[2738]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 24 23:33:49.058276 kubelet[2738]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 24 23:33:49.058647 kubelet[2738]: I0424 23:33:49.058309 2738 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 24 23:33:49.063257 kubelet[2738]: I0424 23:33:49.063239 2738 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 24 23:33:49.063946 kubelet[2738]: I0424 23:33:49.063338 2738 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 24 23:33:49.063946 kubelet[2738]: I0424 23:33:49.063515 2738 server.go:956] "Client rotation is on, will bootstrap in background" Apr 24 23:33:49.064900 kubelet[2738]: I0424 23:33:49.064888 2738 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Apr 24 23:33:49.066312 kubelet[2738]: I0424 23:33:49.066300 2738 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 24 23:33:49.068801 kubelet[2738]: E0424 23:33:49.068367 2738 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 24 23:33:49.068801 kubelet[2738]: I0424 23:33:49.068396 2738 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 24 23:33:49.071875 kubelet[2738]: I0424 23:33:49.071849 2738 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 24 23:33:49.072345 kubelet[2738]: I0424 23:33:49.072320 2738 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 24 23:33:49.072439 kubelet[2738]: I0424 23:33:49.072341 2738 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-6-n-e88c017743","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Apr 24 23:33:49.072502 kubelet[2738]: I0424 23:33:49.072441 2738 topology_manager.go:138] "Creating topology manager with none policy" Apr 24 23:33:49.072502 kubelet[2738]: I0424 23:33:49.072448 2738 container_manager_linux.go:303] "Creating device plugin manager" Apr 24 23:33:49.072502 kubelet[2738]: I0424 23:33:49.072494 2738 state_mem.go:36] "Initialized new in-memory state store" Apr 24 23:33:49.072658 kubelet[2738]: I0424 23:33:49.072634 2738 kubelet.go:480] "Attempting to sync node with API server" Apr 24 23:33:49.072658 kubelet[2738]: I0424 23:33:49.072653 2738 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 24 23:33:49.072693 kubelet[2738]: I0424 23:33:49.072674 2738 kubelet.go:386] "Adding apiserver pod source" Apr 24 23:33:49.072693 kubelet[2738]: I0424 23:33:49.072687 2738 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 24 23:33:49.074077 kubelet[2738]: I0424 23:33:49.074063 2738 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 24 23:33:49.074416 kubelet[2738]: I0424 23:33:49.074402 2738 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 24 23:33:49.076805 kubelet[2738]: I0424 23:33:49.076795 2738 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 24 23:33:49.077296 kubelet[2738]: I0424 23:33:49.077262 2738 server.go:1289] "Started kubelet" Apr 24 23:33:49.083698 kubelet[2738]: I0424 23:33:49.083532 2738 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 24 23:33:49.093227 kubelet[2738]: I0424 23:33:49.093189 2738 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 24 23:33:49.094201 kubelet[2738]: I0424 23:33:49.094174 2738 server.go:317] "Adding debug handlers to kubelet server" Apr 24 23:33:49.098206 kubelet[2738]: I0424 23:33:49.095845 2738 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 24 23:33:49.098716 kubelet[2738]: I0424 23:33:49.098706 2738 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 24 23:33:49.100305 kubelet[2738]: E0424 23:33:49.100292 2738 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-e88c017743\" not found" Apr 24 23:33:49.103077 kubelet[2738]: I0424 23:33:49.103066 2738 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 24 23:33:49.103303 kubelet[2738]: I0424 23:33:49.103296 2738 reconciler.go:26] "Reconciler: start to sync state" Apr 24 23:33:49.104164 kubelet[2738]: I0424 23:33:49.104108 2738 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 24 23:33:49.104390 kubelet[2738]: I0424 23:33:49.104380 2738 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 24 23:33:49.112952 kubelet[2738]: E0424 23:33:49.112929 2738 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 24 23:33:49.113731 kubelet[2738]: I0424 23:33:49.113708 2738 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 24 23:33:49.114489 kubelet[2738]: I0424 23:33:49.114476 2738 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 24 23:33:49.115003 kubelet[2738]: I0424 23:33:49.114993 2738 factory.go:223] Registration of the containerd container factory successfully Apr 24 23:33:49.115073 kubelet[2738]: I0424 23:33:49.115067 2738 factory.go:223] Registration of the systemd container factory successfully Apr 24 23:33:49.115719 kubelet[2738]: I0424 23:33:49.115704 2738 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 24 23:33:49.116128 kubelet[2738]: I0424 23:33:49.116121 2738 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 24 23:33:49.116177 kubelet[2738]: I0424 23:33:49.116171 2738 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 24 23:33:49.116213 kubelet[2738]: I0424 23:33:49.116207 2738 kubelet.go:2436] "Starting kubelet main sync loop" Apr 24 23:33:49.116270 kubelet[2738]: E0424 23:33:49.116260 2738 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 24 23:33:49.171344 kubelet[2738]: I0424 23:33:49.171325 2738 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 24 23:33:49.171449 kubelet[2738]: I0424 23:33:49.171441 2738 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 24 23:33:49.171496 kubelet[2738]: I0424 23:33:49.171491 2738 state_mem.go:36] "Initialized new in-memory state store" Apr 24 23:33:49.171626 kubelet[2738]: I0424 23:33:49.171617 2738 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 24 23:33:49.171665 kubelet[2738]: I0424 23:33:49.171652 2738 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 24 23:33:49.171703 kubelet[2738]: I0424 23:33:49.171697 2738 policy_none.go:49] "None policy: Start" Apr 24 23:33:49.171731 kubelet[2738]: I0424 23:33:49.171726 2738 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 24 23:33:49.171759 kubelet[2738]: I0424 23:33:49.171754 2738 state_mem.go:35] "Initializing new in-memory state store" Apr 24 23:33:49.171848 kubelet[2738]: I0424 23:33:49.171842 2738 state_mem.go:75] "Updated machine memory state" Apr 24 23:33:49.172983 kubelet[2738]: E0424 23:33:49.172970 2738 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 24 23:33:49.174574 kubelet[2738]: I0424 23:33:49.174564 2738 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 24 23:33:49.174642 kubelet[2738]: I0424 23:33:49.174625 2738 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 24 23:33:49.174825 kubelet[2738]: I0424 23:33:49.174817 2738 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 24 23:33:49.175636 kubelet[2738]: E0424 23:33:49.175620 2738 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 24 23:33:49.217759 kubelet[2738]: I0424 23:33:49.217709 2738 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.217928 kubelet[2738]: I0424 23:33:49.217732 2738 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.218036 kubelet[2738]: I0424 23:33:49.218026 2738 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.224100 kubelet[2738]: E0424 23:33:49.224049 2738 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-e88c017743\" already exists" pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.224488 kubelet[2738]: E0424 23:33:49.224419 2738 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-6-n-e88c017743\" already exists" pod="kube-system/kube-scheduler-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.277801 kubelet[2738]: I0424 23:33:49.277756 2738 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.286638 kubelet[2738]: I0424 23:33:49.286566 2738 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.286758 kubelet[2738]: I0424 23:33:49.286659 2738 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305128 kubelet[2738]: I0424 23:33:49.305090 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305128 kubelet[2738]: I0424 23:33:49.305117 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9e69cda47ba3742cdfd1088ebeda4528-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-6-n-e88c017743\" (UID: \"9e69cda47ba3742cdfd1088ebeda4528\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305128 kubelet[2738]: I0424 23:33:49.305131 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-ca-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305128 kubelet[2738]: I0424 23:33:49.305142 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305414 kubelet[2738]: I0424 23:33:49.305155 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305414 kubelet[2738]: I0424 23:33:49.305165 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0fb18c868bc1dc81e6398608291507ac-kubeconfig\") pod \"kube-scheduler-ci-4081-3-6-n-e88c017743\" (UID: \"0fb18c868bc1dc81e6398608291507ac\") " pod="kube-system/kube-scheduler-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305414 kubelet[2738]: I0424 23:33:49.305177 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9e69cda47ba3742cdfd1088ebeda4528-ca-certs\") pod \"kube-apiserver-ci-4081-3-6-n-e88c017743\" (UID: \"9e69cda47ba3742cdfd1088ebeda4528\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305414 kubelet[2738]: I0424 23:33:49.305186 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9e69cda47ba3742cdfd1088ebeda4528-k8s-certs\") pod \"kube-apiserver-ci-4081-3-6-n-e88c017743\" (UID: \"9e69cda47ba3742cdfd1088ebeda4528\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" Apr 24 23:33:49.305414 kubelet[2738]: I0424 23:33:49.305199 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5fd927beb9b34649f92f550ce577c85b-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-6-n-e88c017743\" (UID: \"5fd927beb9b34649f92f550ce577c85b\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" Apr 24 23:33:50.073195 kubelet[2738]: I0424 23:33:50.073162 2738 apiserver.go:52] "Watching apiserver" Apr 24 23:33:50.104154 kubelet[2738]: I0424 23:33:50.104108 2738 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 24 23:33:50.167850 kubelet[2738]: I0424 23:33:50.167790 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-e88c017743" podStartSLOduration=1.167775377 podStartE2EDuration="1.167775377s" podCreationTimestamp="2026-04-24 23:33:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-24 23:33:50.166937586 +0000 UTC m=+1.139088930" watchObservedRunningTime="2026-04-24 23:33:50.167775377 +0000 UTC m=+1.139926711" Apr 24 23:33:50.182307 kubelet[2738]: I0424 23:33:50.182240 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" podStartSLOduration=3.182226279 podStartE2EDuration="3.182226279s" podCreationTimestamp="2026-04-24 23:33:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-24 23:33:50.175315943 +0000 UTC m=+1.147467277" watchObservedRunningTime="2026-04-24 23:33:50.182226279 +0000 UTC m=+1.154377613" Apr 24 23:33:50.182440 kubelet[2738]: I0424 23:33:50.182380 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081-3-6-n-e88c017743" podStartSLOduration=3.182376869 podStartE2EDuration="3.182376869s" podCreationTimestamp="2026-04-24 23:33:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-24 23:33:50.181857979 +0000 UTC m=+1.154009333" watchObservedRunningTime="2026-04-24 23:33:50.182376869 +0000 UTC m=+1.154528213" Apr 24 23:33:53.460642 kubelet[2738]: I0424 23:33:53.459949 2738 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 24 23:33:53.461432 containerd[1631]: time="2026-04-24T23:33:53.460503420Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 24 23:33:53.461934 kubelet[2738]: I0424 23:33:53.460786 2738 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 24 23:33:54.439666 kubelet[2738]: I0424 23:33:54.436653 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4-xtables-lock\") pod \"kube-proxy-c9smv\" (UID: \"65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4\") " pod="kube-system/kube-proxy-c9smv" Apr 24 23:33:54.439666 kubelet[2738]: I0424 23:33:54.436720 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj5rz\" (UniqueName: \"kubernetes.io/projected/65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4-kube-api-access-cj5rz\") pod \"kube-proxy-c9smv\" (UID: \"65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4\") " pod="kube-system/kube-proxy-c9smv" Apr 24 23:33:54.439666 kubelet[2738]: I0424 23:33:54.436752 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4-kube-proxy\") pod \"kube-proxy-c9smv\" (UID: \"65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4\") " pod="kube-system/kube-proxy-c9smv" Apr 24 23:33:54.439666 kubelet[2738]: I0424 23:33:54.436776 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4-lib-modules\") pod \"kube-proxy-c9smv\" (UID: \"65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4\") " pod="kube-system/kube-proxy-c9smv" Apr 24 23:33:54.562756 kubelet[2738]: E0424 23:33:54.562710 2738 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Apr 24 23:33:54.562756 kubelet[2738]: E0424 23:33:54.562748 2738 projected.go:194] Error preparing data for projected volume kube-api-access-cj5rz for pod kube-system/kube-proxy-c9smv: configmap "kube-root-ca.crt" not found Apr 24 23:33:54.563166 kubelet[2738]: E0424 23:33:54.562829 2738 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4-kube-api-access-cj5rz podName:65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4 nodeName:}" failed. No retries permitted until 2026-04-24 23:33:55.062793808 +0000 UTC m=+6.034945172 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cj5rz" (UniqueName: "kubernetes.io/projected/65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4-kube-api-access-cj5rz") pod "kube-proxy-c9smv" (UID: "65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4") : configmap "kube-root-ca.crt" not found Apr 24 23:33:54.638603 kubelet[2738]: I0424 23:33:54.638533 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42k86\" (UniqueName: \"kubernetes.io/projected/db1524da-1b6e-4c92-8551-998a8c0db483-kube-api-access-42k86\") pod \"tigera-operator-6bf85f8dd-97vrl\" (UID: \"db1524da-1b6e-4c92-8551-998a8c0db483\") " pod="tigera-operator/tigera-operator-6bf85f8dd-97vrl" Apr 24 23:33:54.638603 kubelet[2738]: I0424 23:33:54.638611 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/db1524da-1b6e-4c92-8551-998a8c0db483-var-lib-calico\") pod \"tigera-operator-6bf85f8dd-97vrl\" (UID: \"db1524da-1b6e-4c92-8551-998a8c0db483\") " pod="tigera-operator/tigera-operator-6bf85f8dd-97vrl" Apr 24 23:33:54.864241 containerd[1631]: time="2026-04-24T23:33:54.864074689Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6bf85f8dd-97vrl,Uid:db1524da-1b6e-4c92-8551-998a8c0db483,Namespace:tigera-operator,Attempt:0,}" Apr 24 23:33:54.910674 containerd[1631]: time="2026-04-24T23:33:54.910313818Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:33:54.910674 containerd[1631]: time="2026-04-24T23:33:54.910406828Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:33:54.910674 containerd[1631]: time="2026-04-24T23:33:54.910434958Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:54.910674 containerd[1631]: time="2026-04-24T23:33:54.910580928Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:54.976601 containerd[1631]: time="2026-04-24T23:33:54.976566063Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6bf85f8dd-97vrl,Uid:db1524da-1b6e-4c92-8551-998a8c0db483,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"4d52a0870a80c56690fea1c02080ca9015dae165b3a40a491b830ba850d7bed3\"" Apr 24 23:33:54.978069 containerd[1631]: time="2026-04-24T23:33:54.977888354Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\"" Apr 24 23:33:55.346627 containerd[1631]: time="2026-04-24T23:33:55.346571251Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-c9smv,Uid:65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4,Namespace:kube-system,Attempt:0,}" Apr 24 23:33:55.383797 containerd[1631]: time="2026-04-24T23:33:55.383590972Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:33:55.383797 containerd[1631]: time="2026-04-24T23:33:55.383674722Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:33:55.383797 containerd[1631]: time="2026-04-24T23:33:55.383739792Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:55.384209 containerd[1631]: time="2026-04-24T23:33:55.384017982Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:33:55.431590 containerd[1631]: time="2026-04-24T23:33:55.431553792Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-c9smv,Uid:65cef9e2-4a3c-4c3a-a41f-dc54c2c809f4,Namespace:kube-system,Attempt:0,} returns sandbox id \"dfaf8dcaee42da35a4469f1e56c8187513420ee6ce7dce35a70ea14bad937e91\"" Apr 24 23:33:55.436554 containerd[1631]: time="2026-04-24T23:33:55.436533216Z" level=info msg="CreateContainer within sandbox \"dfaf8dcaee42da35a4469f1e56c8187513420ee6ce7dce35a70ea14bad937e91\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 24 23:33:55.448245 containerd[1631]: time="2026-04-24T23:33:55.448166106Z" level=info msg="CreateContainer within sandbox \"dfaf8dcaee42da35a4469f1e56c8187513420ee6ce7dce35a70ea14bad937e91\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"adc581aaa7041385f40e7fae9a1d345abe6005e0638fad92341e0799f615c4fc\"" Apr 24 23:33:55.449360 containerd[1631]: time="2026-04-24T23:33:55.449256037Z" level=info msg="StartContainer for \"adc581aaa7041385f40e7fae9a1d345abe6005e0638fad92341e0799f615c4fc\"" Apr 24 23:33:55.514145 containerd[1631]: time="2026-04-24T23:33:55.513702130Z" level=info msg="StartContainer for \"adc581aaa7041385f40e7fae9a1d345abe6005e0638fad92341e0799f615c4fc\" returns successfully" Apr 24 23:33:56.182664 kubelet[2738]: I0424 23:33:56.182597 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-c9smv" podStartSLOduration=2.182577258 podStartE2EDuration="2.182577258s" podCreationTimestamp="2026-04-24 23:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-24 23:33:56.178690894 +0000 UTC m=+7.150842228" watchObservedRunningTime="2026-04-24 23:33:56.182577258 +0000 UTC m=+7.154728632" Apr 24 23:33:56.508589 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount913984315.mount: Deactivated successfully. Apr 24 23:33:57.037783 containerd[1631]: time="2026-04-24T23:33:57.037735420Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:57.038940 containerd[1631]: time="2026-04-24T23:33:57.038852251Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.7: active requests=0, bytes read=40846156" Apr 24 23:33:57.040622 containerd[1631]: time="2026-04-24T23:33:57.039782192Z" level=info msg="ImageCreate event name:\"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:57.042330 containerd[1631]: time="2026-04-24T23:33:57.041748313Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:33:57.042330 containerd[1631]: time="2026-04-24T23:33:57.042224294Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.7\" with image id \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\", repo tag \"quay.io/tigera/operator:v1.40.7\", repo digest \"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\", size \"40842151\" in 2.06431274s" Apr 24 23:33:57.042330 containerd[1631]: time="2026-04-24T23:33:57.042245264Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\" returns image reference \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\"" Apr 24 23:33:57.046358 containerd[1631]: time="2026-04-24T23:33:57.046317397Z" level=info msg="CreateContainer within sandbox \"4d52a0870a80c56690fea1c02080ca9015dae165b3a40a491b830ba850d7bed3\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Apr 24 23:33:57.056968 containerd[1631]: time="2026-04-24T23:33:57.056932666Z" level=info msg="CreateContainer within sandbox \"4d52a0870a80c56690fea1c02080ca9015dae165b3a40a491b830ba850d7bed3\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1\"" Apr 24 23:33:57.057900 containerd[1631]: time="2026-04-24T23:33:57.057881757Z" level=info msg="StartContainer for \"e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1\"" Apr 24 23:33:57.104363 containerd[1631]: time="2026-04-24T23:33:57.104305075Z" level=info msg="StartContainer for \"e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1\" returns successfully" Apr 24 23:33:57.179553 kubelet[2738]: I0424 23:33:57.179500 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6bf85f8dd-97vrl" podStartSLOduration=1.114070878 podStartE2EDuration="3.179486168s" podCreationTimestamp="2026-04-24 23:33:54 +0000 UTC" firstStartedPulling="2026-04-24 23:33:54.977555294 +0000 UTC m=+5.949706638" lastFinishedPulling="2026-04-24 23:33:57.042970594 +0000 UTC m=+8.015121928" observedRunningTime="2026-04-24 23:33:57.178742617 +0000 UTC m=+8.150893951" watchObservedRunningTime="2026-04-24 23:33:57.179486168 +0000 UTC m=+8.151637502" Apr 24 23:34:02.276282 sudo[1850]: pam_unix(sudo:session): session closed for user root Apr 24 23:34:02.308276 sshd[1846]: pam_unix(sshd:session): session closed for user core Apr 24 23:34:02.314539 systemd[1]: sshd@6-95.217.179.164:22-4.175.71.9:50052.service: Deactivated successfully. Apr 24 23:34:02.322476 systemd[1]: session-7.scope: Deactivated successfully. Apr 24 23:34:02.322522 systemd-logind[1610]: Session 7 logged out. Waiting for processes to exit. Apr 24 23:34:02.325265 systemd-logind[1610]: Removed session 7. Apr 24 23:34:04.204471 kubelet[2738]: I0424 23:34:04.204414 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-cni-log-dir\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.204471 kubelet[2738]: I0424 23:34:04.204442 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-nodeproc\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205322 kubelet[2738]: I0424 23:34:04.204916 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-var-run-calico\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205322 kubelet[2738]: I0424 23:34:04.204939 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-sys-fs\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205322 kubelet[2738]: I0424 23:34:04.204999 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgkvc\" (UniqueName: \"kubernetes.io/projected/0d724d55-d0d0-4971-b7b0-040b8aef1618-kube-api-access-wgkvc\") pod \"calico-typha-84598dc7b-qqml7\" (UID: \"0d724d55-d0d0-4971-b7b0-040b8aef1618\") " pod="calico-system/calico-typha-84598dc7b-qqml7" Apr 24 23:34:04.205322 kubelet[2738]: I0424 23:34:04.205011 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-flexvol-driver-host\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205322 kubelet[2738]: I0424 23:34:04.205022 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-cni-net-dir\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205478 kubelet[2738]: I0424 23:34:04.205032 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-lib-modules\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205478 kubelet[2738]: I0424 23:34:04.205068 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-xtables-lock\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205478 kubelet[2738]: I0424 23:34:04.205078 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-var-lib-calico\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205478 kubelet[2738]: I0424 23:34:04.205135 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/0d724d55-d0d0-4971-b7b0-040b8aef1618-typha-certs\") pod \"calico-typha-84598dc7b-qqml7\" (UID: \"0d724d55-d0d0-4971-b7b0-040b8aef1618\") " pod="calico-system/calico-typha-84598dc7b-qqml7" Apr 24 23:34:04.205478 kubelet[2738]: I0424 23:34:04.205146 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1c0b717-83d4-4824-8da6-05afc154474f-tigera-ca-bundle\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205566 kubelet[2738]: I0424 23:34:04.205174 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/d1c0b717-83d4-4824-8da6-05afc154474f-node-certs\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205566 kubelet[2738]: I0424 23:34:04.205219 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-policysync\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205566 kubelet[2738]: I0424 23:34:04.205233 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d724d55-d0d0-4971-b7b0-040b8aef1618-tigera-ca-bundle\") pod \"calico-typha-84598dc7b-qqml7\" (UID: \"0d724d55-d0d0-4971-b7b0-040b8aef1618\") " pod="calico-system/calico-typha-84598dc7b-qqml7" Apr 24 23:34:04.205566 kubelet[2738]: I0424 23:34:04.205243 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-cni-bin-dir\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205566 kubelet[2738]: I0424 23:34:04.205254 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/d1c0b717-83d4-4824-8da6-05afc154474f-bpffs\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.205638 kubelet[2738]: I0424 23:34:04.205297 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2gfb\" (UniqueName: \"kubernetes.io/projected/d1c0b717-83d4-4824-8da6-05afc154474f-kube-api-access-d2gfb\") pod \"calico-node-6zx6n\" (UID: \"d1c0b717-83d4-4824-8da6-05afc154474f\") " pod="calico-system/calico-node-6zx6n" Apr 24 23:34:04.308132 kubelet[2738]: E0424 23:34:04.306672 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.308132 kubelet[2738]: W0424 23:34:04.306689 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.308132 kubelet[2738]: E0424 23:34:04.306748 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.308132 kubelet[2738]: E0424 23:34:04.307103 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.308132 kubelet[2738]: W0424 23:34:04.307114 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.308132 kubelet[2738]: E0424 23:34:04.307125 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.308132 kubelet[2738]: E0424 23:34:04.307471 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.308132 kubelet[2738]: W0424 23:34:04.307483 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.308132 kubelet[2738]: E0424 23:34:04.307492 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.308132 kubelet[2738]: E0424 23:34:04.307831 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.308959 kubelet[2738]: W0424 23:34:04.307857 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.308959 kubelet[2738]: E0424 23:34:04.307865 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.308959 kubelet[2738]: E0424 23:34:04.308277 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.308959 kubelet[2738]: W0424 23:34:04.308285 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.308959 kubelet[2738]: E0424 23:34:04.308293 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.308959 kubelet[2738]: E0424 23:34:04.308560 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.308959 kubelet[2738]: W0424 23:34:04.308567 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.308959 kubelet[2738]: E0424 23:34:04.308573 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.308959 kubelet[2738]: E0424 23:34:04.308846 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.308959 kubelet[2738]: W0424 23:34:04.308873 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311149 kubelet[2738]: E0424 23:34:04.308881 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311149 kubelet[2738]: E0424 23:34:04.309250 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311149 kubelet[2738]: W0424 23:34:04.309257 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311149 kubelet[2738]: E0424 23:34:04.309265 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311149 kubelet[2738]: E0424 23:34:04.309642 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311149 kubelet[2738]: W0424 23:34:04.309650 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311149 kubelet[2738]: E0424 23:34:04.309657 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311149 kubelet[2738]: E0424 23:34:04.310069 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311149 kubelet[2738]: W0424 23:34:04.310077 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311149 kubelet[2738]: E0424 23:34:04.310096 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311458 kubelet[2738]: E0424 23:34:04.310424 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311458 kubelet[2738]: W0424 23:34:04.310431 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311458 kubelet[2738]: E0424 23:34:04.310440 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311458 kubelet[2738]: E0424 23:34:04.310721 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311458 kubelet[2738]: W0424 23:34:04.310731 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311458 kubelet[2738]: E0424 23:34:04.310738 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311458 kubelet[2738]: E0424 23:34:04.310939 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311458 kubelet[2738]: W0424 23:34:04.310945 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311458 kubelet[2738]: E0424 23:34:04.310951 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311458 kubelet[2738]: E0424 23:34:04.311181 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311752 kubelet[2738]: W0424 23:34:04.311187 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311752 kubelet[2738]: E0424 23:34:04.311194 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311752 kubelet[2738]: E0424 23:34:04.311420 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311752 kubelet[2738]: W0424 23:34:04.311427 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311752 kubelet[2738]: E0424 23:34:04.311433 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311752 kubelet[2738]: E0424 23:34:04.311640 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311752 kubelet[2738]: W0424 23:34:04.311646 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311752 kubelet[2738]: E0424 23:34:04.311652 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.311971 kubelet[2738]: E0424 23:34:04.311848 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.311971 kubelet[2738]: W0424 23:34:04.311855 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.311971 kubelet[2738]: E0424 23:34:04.311861 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.312068 kubelet[2738]: E0424 23:34:04.312056 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.312068 kubelet[2738]: W0424 23:34:04.312063 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.312068 kubelet[2738]: E0424 23:34:04.312069 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.315264 kubelet[2738]: E0424 23:34:04.312820 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.315264 kubelet[2738]: W0424 23:34:04.312830 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.315264 kubelet[2738]: E0424 23:34:04.312966 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.315264 kubelet[2738]: E0424 23:34:04.314074 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.315264 kubelet[2738]: W0424 23:34:04.314109 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.315264 kubelet[2738]: E0424 23:34:04.314118 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.315264 kubelet[2738]: E0424 23:34:04.315206 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.315264 kubelet[2738]: W0424 23:34:04.315215 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.315264 kubelet[2738]: E0424 23:34:04.315226 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.317990 kubelet[2738]: E0424 23:34:04.317967 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.317990 kubelet[2738]: W0424 23:34:04.317985 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.318395 kubelet[2738]: E0424 23:34:04.317997 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.319633 kubelet[2738]: E0424 23:34:04.319506 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.322108 kubelet[2738]: W0424 23:34:04.320951 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.322108 kubelet[2738]: E0424 23:34:04.320972 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.322108 kubelet[2738]: E0424 23:34:04.321139 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:04.322411 kubelet[2738]: E0424 23:34:04.322326 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.322411 kubelet[2738]: W0424 23:34:04.322335 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.322411 kubelet[2738]: E0424 23:34:04.322344 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.325242 kubelet[2738]: E0424 23:34:04.325205 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.325242 kubelet[2738]: W0424 23:34:04.325235 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.325306 kubelet[2738]: E0424 23:34:04.325250 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.325658 kubelet[2738]: E0424 23:34:04.325618 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.325658 kubelet[2738]: W0424 23:34:04.325633 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.325658 kubelet[2738]: E0424 23:34:04.325642 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.326385 kubelet[2738]: E0424 23:34:04.326243 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.326385 kubelet[2738]: W0424 23:34:04.326277 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.326385 kubelet[2738]: E0424 23:34:04.326288 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.332406 kubelet[2738]: E0424 23:34:04.332382 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.332406 kubelet[2738]: W0424 23:34:04.332398 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.332455 kubelet[2738]: E0424 23:34:04.332409 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.333950 kubelet[2738]: E0424 23:34:04.333815 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.333950 kubelet[2738]: W0424 23:34:04.333947 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.334019 kubelet[2738]: E0424 23:34:04.333957 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.335262 kubelet[2738]: E0424 23:34:04.335051 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.335262 kubelet[2738]: W0424 23:34:04.335066 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.335318 kubelet[2738]: E0424 23:34:04.335077 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.336165 kubelet[2738]: E0424 23:34:04.335526 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.336165 kubelet[2738]: W0424 23:34:04.335539 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.336165 kubelet[2738]: E0424 23:34:04.335554 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.336165 kubelet[2738]: E0424 23:34:04.335978 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.336519 kubelet[2738]: W0424 23:34:04.336357 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.336519 kubelet[2738]: E0424 23:34:04.336372 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.337030 kubelet[2738]: E0424 23:34:04.336977 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.337030 kubelet[2738]: W0424 23:34:04.336986 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.337030 kubelet[2738]: E0424 23:34:04.336993 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.340883 kubelet[2738]: E0424 23:34:04.340788 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.340883 kubelet[2738]: W0424 23:34:04.340800 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.340883 kubelet[2738]: E0424 23:34:04.340809 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.341248 kubelet[2738]: E0424 23:34:04.341221 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.341248 kubelet[2738]: W0424 23:34:04.341230 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.341248 kubelet[2738]: E0424 23:34:04.341237 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.346423 kubelet[2738]: E0424 23:34:04.346198 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.346423 kubelet[2738]: W0424 23:34:04.346209 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.346423 kubelet[2738]: E0424 23:34:04.346218 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.347869 kubelet[2738]: E0424 23:34:04.347789 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.347869 kubelet[2738]: W0424 23:34:04.347809 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.347869 kubelet[2738]: E0424 23:34:04.347824 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.349844 kubelet[2738]: E0424 23:34:04.349772 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.349844 kubelet[2738]: W0424 23:34:04.349784 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.349844 kubelet[2738]: E0424 23:34:04.349794 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.352908 kubelet[2738]: E0424 23:34:04.352869 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.352908 kubelet[2738]: W0424 23:34:04.352880 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.352908 kubelet[2738]: E0424 23:34:04.352888 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.359099 kubelet[2738]: E0424 23:34:04.358932 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.359392 kubelet[2738]: W0424 23:34:04.359155 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.359449 kubelet[2738]: E0424 23:34:04.359440 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.360270 kubelet[2738]: E0424 23:34:04.360259 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.360329 kubelet[2738]: W0424 23:34:04.360322 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.361912 kubelet[2738]: E0424 23:34:04.361885 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.362268 kubelet[2738]: E0424 23:34:04.362245 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.362268 kubelet[2738]: W0424 23:34:04.362260 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.362268 kubelet[2738]: E0424 23:34:04.362269 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.363413 kubelet[2738]: E0424 23:34:04.363402 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.363479 kubelet[2738]: W0424 23:34:04.363471 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.363523 kubelet[2738]: E0424 23:34:04.363513 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.364841 kubelet[2738]: E0424 23:34:04.364819 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.364876 kubelet[2738]: W0424 23:34:04.364836 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.364915 kubelet[2738]: E0424 23:34:04.364877 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.365494 kubelet[2738]: E0424 23:34:04.365408 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.365494 kubelet[2738]: W0424 23:34:04.365421 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.365494 kubelet[2738]: E0424 23:34:04.365429 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.365665 kubelet[2738]: E0424 23:34:04.365628 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.365665 kubelet[2738]: W0424 23:34:04.365637 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.365665 kubelet[2738]: E0424 23:34:04.365643 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.373382 kubelet[2738]: E0424 23:34:04.372832 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.373382 kubelet[2738]: W0424 23:34:04.372847 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.373382 kubelet[2738]: E0424 23:34:04.372868 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.376524 kubelet[2738]: E0424 23:34:04.376511 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.376606 kubelet[2738]: W0424 23:34:04.376596 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.376669 kubelet[2738]: E0424 23:34:04.376635 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.377317 kubelet[2738]: E0424 23:34:04.377262 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.377317 kubelet[2738]: W0424 23:34:04.377271 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.377317 kubelet[2738]: E0424 23:34:04.377278 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.406704 kubelet[2738]: E0424 23:34:04.406591 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.406704 kubelet[2738]: W0424 23:34:04.406613 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.406704 kubelet[2738]: E0424 23:34:04.406624 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.406890 kubelet[2738]: E0424 23:34:04.406864 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.406890 kubelet[2738]: W0424 23:34:04.406875 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.406890 kubelet[2738]: E0424 23:34:04.406882 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.407175 kubelet[2738]: E0424 23:34:04.407151 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.407175 kubelet[2738]: W0424 23:34:04.407168 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.407235 kubelet[2738]: E0424 23:34:04.407186 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.407517 kubelet[2738]: E0424 23:34:04.407498 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.407517 kubelet[2738]: W0424 23:34:04.407509 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.407517 kubelet[2738]: E0424 23:34:04.407515 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.407744 kubelet[2738]: E0424 23:34:04.407723 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.407744 kubelet[2738]: W0424 23:34:04.407736 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.407744 kubelet[2738]: E0424 23:34:04.407743 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.407969 kubelet[2738]: E0424 23:34:04.407948 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.407969 kubelet[2738]: W0424 23:34:04.407961 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.407969 kubelet[2738]: E0424 23:34:04.407970 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.408208 kubelet[2738]: E0424 23:34:04.408199 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.408208 kubelet[2738]: W0424 23:34:04.408207 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.408247 kubelet[2738]: E0424 23:34:04.408214 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.408458 kubelet[2738]: E0424 23:34:04.408440 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.408458 kubelet[2738]: W0424 23:34:04.408451 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.408458 kubelet[2738]: E0424 23:34:04.408458 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.408687 kubelet[2738]: E0424 23:34:04.408673 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.408687 kubelet[2738]: W0424 23:34:04.408683 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.408741 kubelet[2738]: E0424 23:34:04.408700 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.408927 kubelet[2738]: E0424 23:34:04.408917 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.408927 kubelet[2738]: W0424 23:34:04.408926 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.408966 kubelet[2738]: E0424 23:34:04.408935 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.409193 kubelet[2738]: E0424 23:34:04.409173 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.409193 kubelet[2738]: W0424 23:34:04.409191 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.409230 kubelet[2738]: E0424 23:34:04.409198 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.409423 kubelet[2738]: E0424 23:34:04.409406 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.409423 kubelet[2738]: W0424 23:34:04.409416 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.409423 kubelet[2738]: E0424 23:34:04.409423 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.409663 kubelet[2738]: E0424 23:34:04.409646 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.409663 kubelet[2738]: W0424 23:34:04.409657 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.409663 kubelet[2738]: E0424 23:34:04.409663 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.409897 kubelet[2738]: E0424 23:34:04.409883 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.409897 kubelet[2738]: W0424 23:34:04.409894 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.409960 kubelet[2738]: E0424 23:34:04.409906 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.410174 kubelet[2738]: E0424 23:34:04.410157 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.410174 kubelet[2738]: W0424 23:34:04.410167 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.410174 kubelet[2738]: E0424 23:34:04.410173 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.410416 kubelet[2738]: E0424 23:34:04.410398 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.410416 kubelet[2738]: W0424 23:34:04.410409 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.410416 kubelet[2738]: E0424 23:34:04.410416 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.410668 kubelet[2738]: E0424 23:34:04.410651 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.410668 kubelet[2738]: W0424 23:34:04.410662 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.410668 kubelet[2738]: E0424 23:34:04.410668 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.410940 kubelet[2738]: E0424 23:34:04.410922 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.410940 kubelet[2738]: W0424 23:34:04.410932 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.410940 kubelet[2738]: E0424 23:34:04.410939 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.411162 kubelet[2738]: E0424 23:34:04.411145 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.411162 kubelet[2738]: W0424 23:34:04.411155 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.411162 kubelet[2738]: E0424 23:34:04.411160 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.411382 kubelet[2738]: E0424 23:34:04.411349 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.411382 kubelet[2738]: W0424 23:34:04.411358 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.411382 kubelet[2738]: E0424 23:34:04.411364 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.419559 containerd[1631]: time="2026-04-24T23:34:04.419523390Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-84598dc7b-qqml7,Uid:0d724d55-d0d0-4971-b7b0-040b8aef1618,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:04.433431 kubelet[2738]: E0424 23:34:04.433407 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.433431 kubelet[2738]: W0424 23:34:04.433424 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.433431 kubelet[2738]: E0424 23:34:04.433435 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.433556 kubelet[2738]: I0424 23:34:04.433457 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7556f2ac-97a1-4b30-8823-9a8d6f3243fe-socket-dir\") pod \"csi-node-driver-kflpl\" (UID: \"7556f2ac-97a1-4b30-8823-9a8d6f3243fe\") " pod="calico-system/csi-node-driver-kflpl" Apr 24 23:34:04.434169 kubelet[2738]: E0424 23:34:04.434149 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.434169 kubelet[2738]: W0424 23:34:04.434162 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.434213 kubelet[2738]: E0424 23:34:04.434170 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.434213 kubelet[2738]: I0424 23:34:04.434197 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/7556f2ac-97a1-4b30-8823-9a8d6f3243fe-varrun\") pod \"csi-node-driver-kflpl\" (UID: \"7556f2ac-97a1-4b30-8823-9a8d6f3243fe\") " pod="calico-system/csi-node-driver-kflpl" Apr 24 23:34:04.434466 kubelet[2738]: E0424 23:34:04.434448 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.434466 kubelet[2738]: W0424 23:34:04.434459 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.434466 kubelet[2738]: E0424 23:34:04.434467 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.434540 kubelet[2738]: I0424 23:34:04.434482 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85bff\" (UniqueName: \"kubernetes.io/projected/7556f2ac-97a1-4b30-8823-9a8d6f3243fe-kube-api-access-85bff\") pod \"csi-node-driver-kflpl\" (UID: \"7556f2ac-97a1-4b30-8823-9a8d6f3243fe\") " pod="calico-system/csi-node-driver-kflpl" Apr 24 23:34:04.434760 kubelet[2738]: E0424 23:34:04.434715 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.434760 kubelet[2738]: W0424 23:34:04.434750 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.434760 kubelet[2738]: E0424 23:34:04.434758 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.434814 kubelet[2738]: I0424 23:34:04.434773 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7556f2ac-97a1-4b30-8823-9a8d6f3243fe-kubelet-dir\") pod \"csi-node-driver-kflpl\" (UID: \"7556f2ac-97a1-4b30-8823-9a8d6f3243fe\") " pod="calico-system/csi-node-driver-kflpl" Apr 24 23:34:04.435062 kubelet[2738]: E0424 23:34:04.435041 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.435062 kubelet[2738]: W0424 23:34:04.435052 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.435062 kubelet[2738]: E0424 23:34:04.435059 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.435551 kubelet[2738]: I0424 23:34:04.435529 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7556f2ac-97a1-4b30-8823-9a8d6f3243fe-registration-dir\") pod \"csi-node-driver-kflpl\" (UID: \"7556f2ac-97a1-4b30-8823-9a8d6f3243fe\") " pod="calico-system/csi-node-driver-kflpl" Apr 24 23:34:04.435639 kubelet[2738]: E0424 23:34:04.435617 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.435639 kubelet[2738]: W0424 23:34:04.435626 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.435639 kubelet[2738]: E0424 23:34:04.435633 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.435874 kubelet[2738]: E0424 23:34:04.435857 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.435874 kubelet[2738]: W0424 23:34:04.435867 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.435874 kubelet[2738]: E0424 23:34:04.435875 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.436390 kubelet[2738]: E0424 23:34:04.436371 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.436390 kubelet[2738]: W0424 23:34:04.436383 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.436390 kubelet[2738]: E0424 23:34:04.436390 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.436650 kubelet[2738]: E0424 23:34:04.436628 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.436650 kubelet[2738]: W0424 23:34:04.436643 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.436714 kubelet[2738]: E0424 23:34:04.436652 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.437196 kubelet[2738]: E0424 23:34:04.437168 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.437196 kubelet[2738]: W0424 23:34:04.437182 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.437196 kubelet[2738]: E0424 23:34:04.437192 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.438027 kubelet[2738]: E0424 23:34:04.437920 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.438027 kubelet[2738]: W0424 23:34:04.437929 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.438027 kubelet[2738]: E0424 23:34:04.437937 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.438974 kubelet[2738]: E0424 23:34:04.438230 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.438974 kubelet[2738]: W0424 23:34:04.438243 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.438974 kubelet[2738]: E0424 23:34:04.438251 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.438974 kubelet[2738]: E0424 23:34:04.438565 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.438974 kubelet[2738]: W0424 23:34:04.438578 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.438974 kubelet[2738]: E0424 23:34:04.438587 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.438974 kubelet[2738]: E0424 23:34:04.438923 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.438974 kubelet[2738]: W0424 23:34:04.438932 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.438974 kubelet[2738]: E0424 23:34:04.438940 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.439568 kubelet[2738]: E0424 23:34:04.439230 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.439568 kubelet[2738]: W0424 23:34:04.439239 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.439568 kubelet[2738]: E0424 23:34:04.439293 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.445641 containerd[1631]: time="2026-04-24T23:34:04.445371602Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:04.445641 containerd[1631]: time="2026-04-24T23:34:04.445417981Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:04.445641 containerd[1631]: time="2026-04-24T23:34:04.445433250Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:04.445641 containerd[1631]: time="2026-04-24T23:34:04.445559048Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:04.493962 containerd[1631]: time="2026-04-24T23:34:04.493124265Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6zx6n,Uid:d1c0b717-83d4-4824-8da6-05afc154474f,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:04.497156 containerd[1631]: time="2026-04-24T23:34:04.497076037Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-84598dc7b-qqml7,Uid:0d724d55-d0d0-4971-b7b0-040b8aef1618,Namespace:calico-system,Attempt:0,} returns sandbox id \"6b931a8c0c8ad8164dc8f5234aa3fd7764d5c64280a8c58f43aa92f36885bfdd\"" Apr 24 23:34:04.500287 containerd[1631]: time="2026-04-24T23:34:04.500263105Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\"" Apr 24 23:34:04.519502 containerd[1631]: time="2026-04-24T23:34:04.519420189Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:04.519613 containerd[1631]: time="2026-04-24T23:34:04.519490787Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:04.519613 containerd[1631]: time="2026-04-24T23:34:04.519503537Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:04.519673 containerd[1631]: time="2026-04-24T23:34:04.519616375Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:04.536065 kubelet[2738]: E0424 23:34:04.536044 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.536473 kubelet[2738]: W0424 23:34:04.536462 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.536537 kubelet[2738]: E0424 23:34:04.536527 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.537779 kubelet[2738]: E0424 23:34:04.537767 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.538736 kubelet[2738]: W0424 23:34:04.538719 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.538803 kubelet[2738]: E0424 23:34:04.538793 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.539245 kubelet[2738]: E0424 23:34:04.539150 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.539809 kubelet[2738]: W0424 23:34:04.539785 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.539855 kubelet[2738]: E0424 23:34:04.539815 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.540174 kubelet[2738]: E0424 23:34:04.540157 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.540174 kubelet[2738]: W0424 23:34:04.540170 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.540207 kubelet[2738]: E0424 23:34:04.540181 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.540551 kubelet[2738]: E0424 23:34:04.540381 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.540551 kubelet[2738]: W0424 23:34:04.540393 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.540551 kubelet[2738]: E0424 23:34:04.540400 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.541186 kubelet[2738]: E0424 23:34:04.541077 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.541186 kubelet[2738]: W0424 23:34:04.541117 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.541186 kubelet[2738]: E0424 23:34:04.541125 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.541828 kubelet[2738]: E0424 23:34:04.541804 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.541828 kubelet[2738]: W0424 23:34:04.541818 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.541828 kubelet[2738]: E0424 23:34:04.541825 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.542106 kubelet[2738]: E0424 23:34:04.542039 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.542106 kubelet[2738]: W0424 23:34:04.542046 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.542106 kubelet[2738]: E0424 23:34:04.542052 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.542278 kubelet[2738]: E0424 23:34:04.542267 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.542278 kubelet[2738]: W0424 23:34:04.542273 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.542278 kubelet[2738]: E0424 23:34:04.542279 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.542499 kubelet[2738]: E0424 23:34:04.542474 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.542499 kubelet[2738]: W0424 23:34:04.542486 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.542499 kubelet[2738]: E0424 23:34:04.542493 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.542724 kubelet[2738]: E0424 23:34:04.542706 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.542724 kubelet[2738]: W0424 23:34:04.542718 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.542724 kubelet[2738]: E0424 23:34:04.542725 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.542962 kubelet[2738]: E0424 23:34:04.542929 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.542962 kubelet[2738]: W0424 23:34:04.542938 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.542962 kubelet[2738]: E0424 23:34:04.542945 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.543202 kubelet[2738]: E0424 23:34:04.543179 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.543202 kubelet[2738]: W0424 23:34:04.543188 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.543202 kubelet[2738]: E0424 23:34:04.543194 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.543461 kubelet[2738]: E0424 23:34:04.543403 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.543461 kubelet[2738]: W0424 23:34:04.543412 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.543461 kubelet[2738]: E0424 23:34:04.543418 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.543713 kubelet[2738]: E0424 23:34:04.543588 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.543713 kubelet[2738]: W0424 23:34:04.543596 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.543713 kubelet[2738]: E0424 23:34:04.543606 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.544497 kubelet[2738]: E0424 23:34:04.543804 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.544497 kubelet[2738]: W0424 23:34:04.543814 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.544497 kubelet[2738]: E0424 23:34:04.543820 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.544497 kubelet[2738]: E0424 23:34:04.544023 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.544497 kubelet[2738]: W0424 23:34:04.544029 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.544497 kubelet[2738]: E0424 23:34:04.544035 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.544497 kubelet[2738]: E0424 23:34:04.544350 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.544497 kubelet[2738]: W0424 23:34:04.544357 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.544497 kubelet[2738]: E0424 23:34:04.544363 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.545285 kubelet[2738]: E0424 23:34:04.545273 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.545393 kubelet[2738]: W0424 23:34:04.545368 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.545463 kubelet[2738]: E0424 23:34:04.545454 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.545810 kubelet[2738]: E0424 23:34:04.545732 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.545810 kubelet[2738]: W0424 23:34:04.545740 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.545810 kubelet[2738]: E0424 23:34:04.545747 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.546210 kubelet[2738]: E0424 23:34:04.546200 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.546811 kubelet[2738]: W0424 23:34:04.546742 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.546811 kubelet[2738]: E0424 23:34:04.546757 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.547206 kubelet[2738]: E0424 23:34:04.547074 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.547206 kubelet[2738]: W0424 23:34:04.547151 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.547206 kubelet[2738]: E0424 23:34:04.547159 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.547592 kubelet[2738]: E0424 23:34:04.547568 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.547641 kubelet[2738]: W0424 23:34:04.547633 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.547702 kubelet[2738]: E0424 23:34:04.547687 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.547981 kubelet[2738]: E0424 23:34:04.547973 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.548040 kubelet[2738]: W0424 23:34:04.548033 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.548070 kubelet[2738]: E0424 23:34:04.548062 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.548397 kubelet[2738]: E0424 23:34:04.548389 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.548464 kubelet[2738]: W0424 23:34:04.548427 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.548464 kubelet[2738]: E0424 23:34:04.548451 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.555804 kubelet[2738]: E0424 23:34:04.555759 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:04.555804 kubelet[2738]: W0424 23:34:04.555770 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:04.555804 kubelet[2738]: E0424 23:34:04.555780 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:04.560912 containerd[1631]: time="2026-04-24T23:34:04.560889704Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6zx6n,Uid:d1c0b717-83d4-4824-8da6-05afc154474f,Namespace:calico-system,Attempt:0,} returns sandbox id \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\"" Apr 24 23:34:06.117992 kubelet[2738]: E0424 23:34:06.117257 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:06.189425 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount494628718.mount: Deactivated successfully. Apr 24 23:34:06.561947 containerd[1631]: time="2026-04-24T23:34:06.561905649Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:06.563120 containerd[1631]: time="2026-04-24T23:34:06.562994820Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.4: active requests=0, bytes read=36107596" Apr 24 23:34:06.564114 containerd[1631]: time="2026-04-24T23:34:06.564069832Z" level=info msg="ImageCreate event name:\"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:06.566308 containerd[1631]: time="2026-04-24T23:34:06.566154987Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:06.567008 containerd[1631]: time="2026-04-24T23:34:06.566589029Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.4\" with image id \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\", size \"36107450\" in 2.066301425s" Apr 24 23:34:06.567008 containerd[1631]: time="2026-04-24T23:34:06.566636438Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\" returns image reference \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\"" Apr 24 23:34:06.568070 containerd[1631]: time="2026-04-24T23:34:06.567931106Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\"" Apr 24 23:34:06.578784 containerd[1631]: time="2026-04-24T23:34:06.578750580Z" level=info msg="CreateContainer within sandbox \"6b931a8c0c8ad8164dc8f5234aa3fd7764d5c64280a8c58f43aa92f36885bfdd\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Apr 24 23:34:06.595977 containerd[1631]: time="2026-04-24T23:34:06.595926646Z" level=info msg="CreateContainer within sandbox \"6b931a8c0c8ad8164dc8f5234aa3fd7764d5c64280a8c58f43aa92f36885bfdd\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"cf9fe04bf71af2c683ad4fe9bda65fe3049331909a4053a38d237169adee12a0\"" Apr 24 23:34:06.596440 containerd[1631]: time="2026-04-24T23:34:06.596422277Z" level=info msg="StartContainer for \"cf9fe04bf71af2c683ad4fe9bda65fe3049331909a4053a38d237169adee12a0\"" Apr 24 23:34:06.662573 containerd[1631]: time="2026-04-24T23:34:06.662438975Z" level=info msg="StartContainer for \"cf9fe04bf71af2c683ad4fe9bda65fe3049331909a4053a38d237169adee12a0\" returns successfully" Apr 24 23:34:07.213774 kubelet[2738]: I0424 23:34:07.213662 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-84598dc7b-qqml7" podStartSLOduration=1.146361203 podStartE2EDuration="3.21364375s" podCreationTimestamp="2026-04-24 23:34:04 +0000 UTC" firstStartedPulling="2026-04-24 23:34:04.499828383 +0000 UTC m=+15.471979727" lastFinishedPulling="2026-04-24 23:34:06.56711093 +0000 UTC m=+17.539262274" observedRunningTime="2026-04-24 23:34:07.212159534 +0000 UTC m=+18.184310918" watchObservedRunningTime="2026-04-24 23:34:07.21364375 +0000 UTC m=+18.185795134" Apr 24 23:34:07.229302 kubelet[2738]: E0424 23:34:07.229257 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.229302 kubelet[2738]: W0424 23:34:07.229288 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.229302 kubelet[2738]: E0424 23:34:07.229311 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.230145 kubelet[2738]: E0424 23:34:07.229916 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.230145 kubelet[2738]: W0424 23:34:07.229938 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.230145 kubelet[2738]: E0424 23:34:07.229961 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.230735 kubelet[2738]: E0424 23:34:07.230594 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.230735 kubelet[2738]: W0424 23:34:07.230614 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.230735 kubelet[2738]: E0424 23:34:07.230631 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.231489 kubelet[2738]: E0424 23:34:07.231360 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.231489 kubelet[2738]: W0424 23:34:07.231379 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.231489 kubelet[2738]: E0424 23:34:07.231395 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.232570 kubelet[2738]: E0424 23:34:07.232152 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.232570 kubelet[2738]: W0424 23:34:07.232194 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.232570 kubelet[2738]: E0424 23:34:07.232217 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.233032 kubelet[2738]: E0424 23:34:07.232918 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.233032 kubelet[2738]: W0424 23:34:07.232938 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.233032 kubelet[2738]: E0424 23:34:07.232955 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.234009 kubelet[2738]: E0424 23:34:07.233712 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.234009 kubelet[2738]: W0424 23:34:07.233736 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.234009 kubelet[2738]: E0424 23:34:07.233756 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.235259 kubelet[2738]: E0424 23:34:07.235027 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.235259 kubelet[2738]: W0424 23:34:07.235050 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.235259 kubelet[2738]: E0424 23:34:07.235068 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.236876 kubelet[2738]: E0424 23:34:07.236578 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.236876 kubelet[2738]: W0424 23:34:07.236601 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.236876 kubelet[2738]: E0424 23:34:07.236619 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.237759 kubelet[2738]: E0424 23:34:07.237738 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.237862 kubelet[2738]: W0424 23:34:07.237839 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.237862 kubelet[2738]: E0424 23:34:07.237860 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.238609 kubelet[2738]: E0424 23:34:07.238575 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.238609 kubelet[2738]: W0424 23:34:07.238599 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.238745 kubelet[2738]: E0424 23:34:07.238617 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.239188 kubelet[2738]: E0424 23:34:07.239155 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.239188 kubelet[2738]: W0424 23:34:07.239178 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.239297 kubelet[2738]: E0424 23:34:07.239193 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.239869 kubelet[2738]: E0424 23:34:07.239625 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.239869 kubelet[2738]: W0424 23:34:07.239654 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.239869 kubelet[2738]: E0424 23:34:07.239722 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.240275 kubelet[2738]: E0424 23:34:07.240242 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.240275 kubelet[2738]: W0424 23:34:07.240265 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.240433 kubelet[2738]: E0424 23:34:07.240283 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.240810 kubelet[2738]: E0424 23:34:07.240777 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.240810 kubelet[2738]: W0424 23:34:07.240798 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.240913 kubelet[2738]: E0424 23:34:07.240813 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.259589 kubelet[2738]: E0424 23:34:07.259548 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.259589 kubelet[2738]: W0424 23:34:07.259576 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.259746 kubelet[2738]: E0424 23:34:07.259596 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.260295 kubelet[2738]: E0424 23:34:07.260252 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.260295 kubelet[2738]: W0424 23:34:07.260282 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.260479 kubelet[2738]: E0424 23:34:07.260300 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.260794 kubelet[2738]: E0424 23:34:07.260769 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.260890 kubelet[2738]: W0424 23:34:07.260855 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.260890 kubelet[2738]: E0424 23:34:07.260881 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.261420 kubelet[2738]: E0424 23:34:07.261389 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.261420 kubelet[2738]: W0424 23:34:07.261410 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.261540 kubelet[2738]: E0424 23:34:07.261426 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.261957 kubelet[2738]: E0424 23:34:07.261915 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.261957 kubelet[2738]: W0424 23:34:07.261944 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.262212 kubelet[2738]: E0424 23:34:07.261962 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.262711 kubelet[2738]: E0424 23:34:07.262660 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.262711 kubelet[2738]: W0424 23:34:07.262697 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.262890 kubelet[2738]: E0424 23:34:07.262714 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.263300 kubelet[2738]: E0424 23:34:07.263273 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.263300 kubelet[2738]: W0424 23:34:07.263294 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.263430 kubelet[2738]: E0424 23:34:07.263308 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.263782 kubelet[2738]: E0424 23:34:07.263744 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.263870 kubelet[2738]: W0424 23:34:07.263791 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.263870 kubelet[2738]: E0424 23:34:07.263808 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.264289 kubelet[2738]: E0424 23:34:07.264257 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.264289 kubelet[2738]: W0424 23:34:07.264278 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.264372 kubelet[2738]: E0424 23:34:07.264296 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.264740 kubelet[2738]: E0424 23:34:07.264708 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.264740 kubelet[2738]: W0424 23:34:07.264728 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.264843 kubelet[2738]: E0424 23:34:07.264742 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.265305 kubelet[2738]: E0424 23:34:07.265273 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.265305 kubelet[2738]: W0424 23:34:07.265294 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.265391 kubelet[2738]: E0424 23:34:07.265309 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.266268 kubelet[2738]: E0424 23:34:07.266234 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.266268 kubelet[2738]: W0424 23:34:07.266257 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.266373 kubelet[2738]: E0424 23:34:07.266273 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.266769 kubelet[2738]: E0424 23:34:07.266745 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.266866 kubelet[2738]: W0424 23:34:07.266826 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.266866 kubelet[2738]: E0424 23:34:07.266853 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.267381 kubelet[2738]: E0424 23:34:07.267340 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.267381 kubelet[2738]: W0424 23:34:07.267366 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.267381 kubelet[2738]: E0424 23:34:07.267383 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.267926 kubelet[2738]: E0424 23:34:07.267889 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.267926 kubelet[2738]: W0424 23:34:07.267910 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.267926 kubelet[2738]: E0424 23:34:07.267926 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.268661 kubelet[2738]: E0424 23:34:07.268621 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.268661 kubelet[2738]: W0424 23:34:07.268647 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.268661 kubelet[2738]: E0424 23:34:07.268664 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.269509 kubelet[2738]: E0424 23:34:07.269472 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.269509 kubelet[2738]: W0424 23:34:07.269496 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.269631 kubelet[2738]: E0424 23:34:07.269516 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.270034 kubelet[2738]: E0424 23:34:07.269994 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:07.270034 kubelet[2738]: W0424 23:34:07.270019 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:07.270034 kubelet[2738]: E0424 23:34:07.270036 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:07.574760 systemd[1]: run-containerd-runc-k8s.io-cf9fe04bf71af2c683ad4fe9bda65fe3049331909a4053a38d237169adee12a0-runc.uj7shQ.mount: Deactivated successfully. Apr 24 23:34:08.119150 kubelet[2738]: E0424 23:34:08.117357 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:08.200417 kubelet[2738]: I0424 23:34:08.200387 2738 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 24 23:34:08.248475 kubelet[2738]: E0424 23:34:08.248444 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.248475 kubelet[2738]: W0424 23:34:08.248463 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.248475 kubelet[2738]: E0424 23:34:08.248478 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.248906 kubelet[2738]: E0424 23:34:08.248792 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.248906 kubelet[2738]: W0424 23:34:08.248799 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.248906 kubelet[2738]: E0424 23:34:08.248806 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.249061 kubelet[2738]: E0424 23:34:08.248985 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.249061 kubelet[2738]: W0424 23:34:08.249057 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.249142 kubelet[2738]: E0424 23:34:08.249066 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.249696 kubelet[2738]: E0424 23:34:08.249574 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.249696 kubelet[2738]: W0424 23:34:08.249584 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.249696 kubelet[2738]: E0424 23:34:08.249616 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.249955 kubelet[2738]: E0424 23:34:08.249927 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.249955 kubelet[2738]: W0424 23:34:08.249938 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.249955 kubelet[2738]: E0424 23:34:08.249945 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.250297 kubelet[2738]: E0424 23:34:08.250277 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.250297 kubelet[2738]: W0424 23:34:08.250288 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.250384 kubelet[2738]: E0424 23:34:08.250364 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.250754 kubelet[2738]: E0424 23:34:08.250730 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.250754 kubelet[2738]: W0424 23:34:08.250740 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.250754 kubelet[2738]: E0424 23:34:08.250747 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.251109 kubelet[2738]: E0424 23:34:08.251073 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.251109 kubelet[2738]: W0424 23:34:08.251097 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.251151 kubelet[2738]: E0424 23:34:08.251103 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.251424 kubelet[2738]: E0424 23:34:08.251407 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.251424 kubelet[2738]: W0424 23:34:08.251418 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.251424 kubelet[2738]: E0424 23:34:08.251425 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.251700 kubelet[2738]: E0424 23:34:08.251629 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.251700 kubelet[2738]: W0424 23:34:08.251638 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.251700 kubelet[2738]: E0424 23:34:08.251689 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.252073 kubelet[2738]: E0424 23:34:08.251950 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.252073 kubelet[2738]: W0424 23:34:08.251959 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.252073 kubelet[2738]: E0424 23:34:08.251966 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.252386 kubelet[2738]: E0424 23:34:08.252369 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.252386 kubelet[2738]: W0424 23:34:08.252380 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.252386 kubelet[2738]: E0424 23:34:08.252386 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.252788 kubelet[2738]: E0424 23:34:08.252770 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.252788 kubelet[2738]: W0424 23:34:08.252782 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.252788 kubelet[2738]: E0424 23:34:08.252789 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.253311 kubelet[2738]: E0424 23:34:08.253292 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.253311 kubelet[2738]: W0424 23:34:08.253304 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.253311 kubelet[2738]: E0424 23:34:08.253311 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.253623 kubelet[2738]: E0424 23:34:08.253608 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.253623 kubelet[2738]: W0424 23:34:08.253620 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.253666 kubelet[2738]: E0424 23:34:08.253626 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.261846 containerd[1631]: time="2026-04-24T23:34:08.261805799Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:08.262943 containerd[1631]: time="2026-04-24T23:34:08.262760403Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4: active requests=0, bytes read=4630250" Apr 24 23:34:08.264023 containerd[1631]: time="2026-04-24T23:34:08.263982355Z" level=info msg="ImageCreate event name:\"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:08.266049 containerd[1631]: time="2026-04-24T23:34:08.266024225Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:08.266704 containerd[1631]: time="2026-04-24T23:34:08.266618227Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" with image id \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\", size \"6186255\" in 1.697963113s" Apr 24 23:34:08.266704 containerd[1631]: time="2026-04-24T23:34:08.266642406Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" returns image reference \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\"" Apr 24 23:34:08.268983 kubelet[2738]: E0424 23:34:08.268955 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.268983 kubelet[2738]: W0424 23:34:08.268972 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.268983 kubelet[2738]: E0424 23:34:08.268983 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.269783 kubelet[2738]: E0424 23:34:08.269609 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.269783 kubelet[2738]: W0424 23:34:08.269626 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.269783 kubelet[2738]: E0424 23:34:08.269644 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.270113 kubelet[2738]: E0424 23:34:08.270022 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.270113 kubelet[2738]: W0424 23:34:08.270032 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.270113 kubelet[2738]: E0424 23:34:08.270039 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.270377 kubelet[2738]: E0424 23:34:08.270357 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.270655 kubelet[2738]: W0424 23:34:08.270646 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.270780 kubelet[2738]: E0424 23:34:08.270706 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.270879 containerd[1631]: time="2026-04-24T23:34:08.270817963Z" level=info msg="CreateContainer within sandbox \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Apr 24 23:34:08.271418 kubelet[2738]: E0424 23:34:08.271198 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.271418 kubelet[2738]: W0424 23:34:08.271206 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.271878 kubelet[2738]: E0424 23:34:08.271785 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.272126 kubelet[2738]: E0424 23:34:08.272039 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.272126 kubelet[2738]: W0424 23:34:08.272048 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.272126 kubelet[2738]: E0424 23:34:08.272055 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.272418 kubelet[2738]: E0424 23:34:08.272338 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.272418 kubelet[2738]: W0424 23:34:08.272346 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.272418 kubelet[2738]: E0424 23:34:08.272353 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.272618 kubelet[2738]: E0424 23:34:08.272610 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.272725 kubelet[2738]: W0424 23:34:08.272649 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.272725 kubelet[2738]: E0424 23:34:08.272658 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.272921 kubelet[2738]: E0424 23:34:08.272914 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.273109 kubelet[2738]: W0424 23:34:08.272953 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.273109 kubelet[2738]: E0424 23:34:08.272961 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.273436 kubelet[2738]: E0424 23:34:08.273427 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.273495 kubelet[2738]: W0424 23:34:08.273488 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.273529 kubelet[2738]: E0424 23:34:08.273521 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.273821 kubelet[2738]: E0424 23:34:08.273813 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.273884 kubelet[2738]: W0424 23:34:08.273858 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.273884 kubelet[2738]: E0424 23:34:08.273867 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.274258 kubelet[2738]: E0424 23:34:08.274235 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.274258 kubelet[2738]: W0424 23:34:08.274252 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.274322 kubelet[2738]: E0424 23:34:08.274261 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.274730 kubelet[2738]: E0424 23:34:08.274706 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.274730 kubelet[2738]: W0424 23:34:08.274718 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.274801 kubelet[2738]: E0424 23:34:08.274777 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.275191 kubelet[2738]: E0424 23:34:08.275173 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.275191 kubelet[2738]: W0424 23:34:08.275184 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.275239 kubelet[2738]: E0424 23:34:08.275192 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.275741 kubelet[2738]: E0424 23:34:08.275720 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.275741 kubelet[2738]: W0424 23:34:08.275732 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.275741 kubelet[2738]: E0424 23:34:08.275739 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.276021 kubelet[2738]: E0424 23:34:08.276002 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.276050 kubelet[2738]: W0424 23:34:08.276023 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.276050 kubelet[2738]: E0424 23:34:08.276030 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.276580 kubelet[2738]: E0424 23:34:08.276561 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.276580 kubelet[2738]: W0424 23:34:08.276573 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.276580 kubelet[2738]: E0424 23:34:08.276580 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.277063 kubelet[2738]: E0424 23:34:08.277040 2738 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 24 23:34:08.277063 kubelet[2738]: W0424 23:34:08.277053 2738 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 24 23:34:08.277063 kubelet[2738]: E0424 23:34:08.277065 2738 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 24 23:34:08.286138 containerd[1631]: time="2026-04-24T23:34:08.286069655Z" level=info msg="CreateContainer within sandbox \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"00ee8e2006180190772de4881e0db5bb569f91df8588de1fc23832bd35e69850\"" Apr 24 23:34:08.287203 containerd[1631]: time="2026-04-24T23:34:08.286596757Z" level=info msg="StartContainer for \"00ee8e2006180190772de4881e0db5bb569f91df8588de1fc23832bd35e69850\"" Apr 24 23:34:08.339286 containerd[1631]: time="2026-04-24T23:34:08.339176420Z" level=info msg="StartContainer for \"00ee8e2006180190772de4881e0db5bb569f91df8588de1fc23832bd35e69850\" returns successfully" Apr 24 23:34:08.367152 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-00ee8e2006180190772de4881e0db5bb569f91df8588de1fc23832bd35e69850-rootfs.mount: Deactivated successfully. Apr 24 23:34:08.457332 containerd[1631]: time="2026-04-24T23:34:08.457052995Z" level=info msg="shim disconnected" id=00ee8e2006180190772de4881e0db5bb569f91df8588de1fc23832bd35e69850 namespace=k8s.io Apr 24 23:34:08.457332 containerd[1631]: time="2026-04-24T23:34:08.457143133Z" level=warning msg="cleaning up after shim disconnected" id=00ee8e2006180190772de4881e0db5bb569f91df8588de1fc23832bd35e69850 namespace=k8s.io Apr 24 23:34:08.457332 containerd[1631]: time="2026-04-24T23:34:08.457150323Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 23:34:09.207072 containerd[1631]: time="2026-04-24T23:34:09.206878461Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\"" Apr 24 23:34:10.117706 kubelet[2738]: E0424 23:34:10.117174 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:12.056632 update_engine[1614]: I20260424 23:34:12.056574 1614 update_attempter.cc:509] Updating boot flags... Apr 24 23:34:12.109168 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 31 scanned by (udev-worker) (3545) Apr 24 23:34:12.117429 kubelet[2738]: E0424 23:34:12.117303 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:12.183610 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 31 scanned by (udev-worker) (3546) Apr 24 23:34:13.329626 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3158784876.mount: Deactivated successfully. Apr 24 23:34:13.358327 containerd[1631]: time="2026-04-24T23:34:13.358282559Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:13.359467 containerd[1631]: time="2026-04-24T23:34:13.359384237Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.4: active requests=0, bytes read=159838564" Apr 24 23:34:13.361273 containerd[1631]: time="2026-04-24T23:34:13.360297388Z" level=info msg="ImageCreate event name:\"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:13.362763 containerd[1631]: time="2026-04-24T23:34:13.362127939Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:13.362763 containerd[1631]: time="2026-04-24T23:34:13.362663913Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.4\" with image id \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\", size \"159838426\" in 4.155753112s" Apr 24 23:34:13.362763 containerd[1631]: time="2026-04-24T23:34:13.362697882Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\" returns image reference \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\"" Apr 24 23:34:13.365707 containerd[1631]: time="2026-04-24T23:34:13.365657661Z" level=info msg="CreateContainer within sandbox \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Apr 24 23:34:13.381716 containerd[1631]: time="2026-04-24T23:34:13.381665251Z" level=info msg="CreateContainer within sandbox \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"e4b0006fd0507e549477415022696d722d1e51cd0e2e18eb8aac699592a9a9eb\"" Apr 24 23:34:13.382404 containerd[1631]: time="2026-04-24T23:34:13.382233995Z" level=info msg="StartContainer for \"e4b0006fd0507e549477415022696d722d1e51cd0e2e18eb8aac699592a9a9eb\"" Apr 24 23:34:13.413073 systemd[1]: run-containerd-runc-k8s.io-e4b0006fd0507e549477415022696d722d1e51cd0e2e18eb8aac699592a9a9eb-runc.O2sdpq.mount: Deactivated successfully. Apr 24 23:34:13.440586 containerd[1631]: time="2026-04-24T23:34:13.440287729Z" level=info msg="StartContainer for \"e4b0006fd0507e549477415022696d722d1e51cd0e2e18eb8aac699592a9a9eb\" returns successfully" Apr 24 23:34:13.546537 containerd[1631]: time="2026-04-24T23:34:13.546474782Z" level=info msg="shim disconnected" id=e4b0006fd0507e549477415022696d722d1e51cd0e2e18eb8aac699592a9a9eb namespace=k8s.io Apr 24 23:34:13.546537 containerd[1631]: time="2026-04-24T23:34:13.546528871Z" level=warning msg="cleaning up after shim disconnected" id=e4b0006fd0507e549477415022696d722d1e51cd0e2e18eb8aac699592a9a9eb namespace=k8s.io Apr 24 23:34:13.546537 containerd[1631]: time="2026-04-24T23:34:13.546538671Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 23:34:13.557391 containerd[1631]: time="2026-04-24T23:34:13.557347806Z" level=warning msg="cleanup warnings time=\"2026-04-24T23:34:13Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Apr 24 23:34:14.117387 kubelet[2738]: E0424 23:34:14.117292 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:14.218895 containerd[1631]: time="2026-04-24T23:34:14.218833812Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\"" Apr 24 23:34:14.333380 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e4b0006fd0507e549477415022696d722d1e51cd0e2e18eb8aac699592a9a9eb-rootfs.mount: Deactivated successfully. Apr 24 23:34:16.117312 kubelet[2738]: E0424 23:34:16.117265 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:16.738921 containerd[1631]: time="2026-04-24T23:34:16.738883200Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:16.739800 containerd[1631]: time="2026-04-24T23:34:16.739711812Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.4: active requests=0, bytes read=70611671" Apr 24 23:34:16.741542 containerd[1631]: time="2026-04-24T23:34:16.740649845Z" level=info msg="ImageCreate event name:\"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:16.742865 containerd[1631]: time="2026-04-24T23:34:16.742268880Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:16.742865 containerd[1631]: time="2026-04-24T23:34:16.742780116Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.4\" with image id \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\", size \"72167716\" in 2.523901174s" Apr 24 23:34:16.742865 containerd[1631]: time="2026-04-24T23:34:16.742801206Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\" returns image reference \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\"" Apr 24 23:34:16.746372 containerd[1631]: time="2026-04-24T23:34:16.746333746Z" level=info msg="CreateContainer within sandbox \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Apr 24 23:34:16.768664 containerd[1631]: time="2026-04-24T23:34:16.768604494Z" level=info msg="CreateContainer within sandbox \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"28cdace963488417c920d158f00fac8865491525c66046126454a9db8a9bf624\"" Apr 24 23:34:16.770142 containerd[1631]: time="2026-04-24T23:34:16.769025591Z" level=info msg="StartContainer for \"28cdace963488417c920d158f00fac8865491525c66046126454a9db8a9bf624\"" Apr 24 23:34:16.824067 containerd[1631]: time="2026-04-24T23:34:16.824017308Z" level=info msg="StartContainer for \"28cdace963488417c920d158f00fac8865491525c66046126454a9db8a9bf624\" returns successfully" Apr 24 23:34:17.251547 containerd[1631]: time="2026-04-24T23:34:17.251488850Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 24 23:34:17.270897 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-28cdace963488417c920d158f00fac8865491525c66046126454a9db8a9bf624-rootfs.mount: Deactivated successfully. Apr 24 23:34:17.274029 containerd[1631]: time="2026-04-24T23:34:17.273955090Z" level=info msg="shim disconnected" id=28cdace963488417c920d158f00fac8865491525c66046126454a9db8a9bf624 namespace=k8s.io Apr 24 23:34:17.274029 containerd[1631]: time="2026-04-24T23:34:17.274024259Z" level=warning msg="cleaning up after shim disconnected" id=28cdace963488417c920d158f00fac8865491525c66046126454a9db8a9bf624 namespace=k8s.io Apr 24 23:34:17.274029 containerd[1631]: time="2026-04-24T23:34:17.274030829Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 23:34:17.304920 kubelet[2738]: I0424 23:34:17.304816 2738 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Apr 24 23:34:17.431969 kubelet[2738]: I0424 23:34:17.431932 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtmb6\" (UniqueName: \"kubernetes.io/projected/51e78ef8-95fe-4fb0-adbe-918fd6d5713b-kube-api-access-mtmb6\") pod \"coredns-674b8bbfcf-kc2pn\" (UID: \"51e78ef8-95fe-4fb0-adbe-918fd6d5713b\") " pod="kube-system/coredns-674b8bbfcf-kc2pn" Apr 24 23:34:17.432198 kubelet[2738]: I0424 23:34:17.431986 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f187502e-fb21-4086-9792-3a608142290e-goldmane-ca-bundle\") pod \"goldmane-5b85766d88-l624f\" (UID: \"f187502e-fb21-4086-9792-3a608142290e\") " pod="calico-system/goldmane-5b85766d88-l624f" Apr 24 23:34:17.432198 kubelet[2738]: I0424 23:34:17.432006 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsdc4\" (UniqueName: \"kubernetes.io/projected/5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b-kube-api-access-nsdc4\") pod \"calico-kube-controllers-6cd754fd7f-hzjqr\" (UID: \"5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b\") " pod="calico-system/calico-kube-controllers-6cd754fd7f-hzjqr" Apr 24 23:34:17.432198 kubelet[2738]: I0424 23:34:17.432033 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51e78ef8-95fe-4fb0-adbe-918fd6d5713b-config-volume\") pod \"coredns-674b8bbfcf-kc2pn\" (UID: \"51e78ef8-95fe-4fb0-adbe-918fd6d5713b\") " pod="kube-system/coredns-674b8bbfcf-kc2pn" Apr 24 23:34:17.432198 kubelet[2738]: I0424 23:34:17.432051 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tw6n\" (UniqueName: \"kubernetes.io/projected/fd07f2b5-728d-4d3c-a1de-b9db458623b8-kube-api-access-4tw6n\") pod \"whisker-54cb6594cd-g4bq8\" (UID: \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\") " pod="calico-system/whisker-54cb6594cd-g4bq8" Apr 24 23:34:17.432198 kubelet[2738]: I0424 23:34:17.432063 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtzm7\" (UniqueName: \"kubernetes.io/projected/f187502e-fb21-4086-9792-3a608142290e-kube-api-access-gtzm7\") pod \"goldmane-5b85766d88-l624f\" (UID: \"f187502e-fb21-4086-9792-3a608142290e\") " pod="calico-system/goldmane-5b85766d88-l624f" Apr 24 23:34:17.432304 kubelet[2738]: I0424 23:34:17.432075 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/fd07f2b5-728d-4d3c-a1de-b9db458623b8-nginx-config\") pod \"whisker-54cb6594cd-g4bq8\" (UID: \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\") " pod="calico-system/whisker-54cb6594cd-g4bq8" Apr 24 23:34:17.432304 kubelet[2738]: I0424 23:34:17.432110 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/fd07f2b5-728d-4d3c-a1de-b9db458623b8-whisker-backend-key-pair\") pod \"whisker-54cb6594cd-g4bq8\" (UID: \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\") " pod="calico-system/whisker-54cb6594cd-g4bq8" Apr 24 23:34:17.432304 kubelet[2738]: I0424 23:34:17.432130 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b-calico-apiserver-certs\") pod \"calico-apiserver-85f77d67c7-7k5mt\" (UID: \"c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b\") " pod="calico-system/calico-apiserver-85f77d67c7-7k5mt" Apr 24 23:34:17.432304 kubelet[2738]: I0424 23:34:17.432147 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/f187502e-fb21-4086-9792-3a608142290e-goldmane-key-pair\") pod \"goldmane-5b85766d88-l624f\" (UID: \"f187502e-fb21-4086-9792-3a608142290e\") " pod="calico-system/goldmane-5b85766d88-l624f" Apr 24 23:34:17.432304 kubelet[2738]: I0424 23:34:17.432167 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwz25\" (UniqueName: \"kubernetes.io/projected/c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b-kube-api-access-lwz25\") pod \"calico-apiserver-85f77d67c7-7k5mt\" (UID: \"c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b\") " pod="calico-system/calico-apiserver-85f77d67c7-7k5mt" Apr 24 23:34:17.432393 kubelet[2738]: I0424 23:34:17.432185 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b-tigera-ca-bundle\") pod \"calico-kube-controllers-6cd754fd7f-hzjqr\" (UID: \"5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b\") " pod="calico-system/calico-kube-controllers-6cd754fd7f-hzjqr" Apr 24 23:34:17.432393 kubelet[2738]: I0424 23:34:17.432199 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f187502e-fb21-4086-9792-3a608142290e-config\") pod \"goldmane-5b85766d88-l624f\" (UID: \"f187502e-fb21-4086-9792-3a608142290e\") " pod="calico-system/goldmane-5b85766d88-l624f" Apr 24 23:34:17.432393 kubelet[2738]: I0424 23:34:17.432210 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/0887e740-8164-4423-b084-a17d88aff0ef-calico-apiserver-certs\") pod \"calico-apiserver-85f77d67c7-lz87d\" (UID: \"0887e740-8164-4423-b084-a17d88aff0ef\") " pod="calico-system/calico-apiserver-85f77d67c7-lz87d" Apr 24 23:34:17.432393 kubelet[2738]: I0424 23:34:17.432229 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvnsm\" (UniqueName: \"kubernetes.io/projected/0887e740-8164-4423-b084-a17d88aff0ef-kube-api-access-jvnsm\") pod \"calico-apiserver-85f77d67c7-lz87d\" (UID: \"0887e740-8164-4423-b084-a17d88aff0ef\") " pod="calico-system/calico-apiserver-85f77d67c7-lz87d" Apr 24 23:34:17.432393 kubelet[2738]: I0424 23:34:17.432240 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8da1862-fd17-4f27-a872-feaa69b7162e-config-volume\") pod \"coredns-674b8bbfcf-2gd6v\" (UID: \"d8da1862-fd17-4f27-a872-feaa69b7162e\") " pod="kube-system/coredns-674b8bbfcf-2gd6v" Apr 24 23:34:17.432481 kubelet[2738]: I0424 23:34:17.432256 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5snf8\" (UniqueName: \"kubernetes.io/projected/d8da1862-fd17-4f27-a872-feaa69b7162e-kube-api-access-5snf8\") pod \"coredns-674b8bbfcf-2gd6v\" (UID: \"d8da1862-fd17-4f27-a872-feaa69b7162e\") " pod="kube-system/coredns-674b8bbfcf-2gd6v" Apr 24 23:34:17.432481 kubelet[2738]: I0424 23:34:17.432273 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd07f2b5-728d-4d3c-a1de-b9db458623b8-whisker-ca-bundle\") pod \"whisker-54cb6594cd-g4bq8\" (UID: \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\") " pod="calico-system/whisker-54cb6594cd-g4bq8" Apr 24 23:34:17.642682 containerd[1631]: time="2026-04-24T23:34:17.642611887Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-kc2pn,Uid:51e78ef8-95fe-4fb0-adbe-918fd6d5713b,Namespace:kube-system,Attempt:0,}" Apr 24 23:34:17.645259 containerd[1631]: time="2026-04-24T23:34:17.645156198Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-54cb6594cd-g4bq8,Uid:fd07f2b5-728d-4d3c-a1de-b9db458623b8,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:17.654271 containerd[1631]: time="2026-04-24T23:34:17.654237244Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-l624f,Uid:f187502e-fb21-4086-9792-3a608142290e,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:17.655835 containerd[1631]: time="2026-04-24T23:34:17.655808942Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cd754fd7f-hzjqr,Uid:5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:17.668489 containerd[1631]: time="2026-04-24T23:34:17.668263592Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85f77d67c7-7k5mt,Uid:c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:17.668489 containerd[1631]: time="2026-04-24T23:34:17.668408211Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2gd6v,Uid:d8da1862-fd17-4f27-a872-feaa69b7162e,Namespace:kube-system,Attempt:0,}" Apr 24 23:34:17.669993 containerd[1631]: time="2026-04-24T23:34:17.669976778Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85f77d67c7-lz87d,Uid:0887e740-8164-4423-b084-a17d88aff0ef,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:17.814704 containerd[1631]: time="2026-04-24T23:34:17.814612920Z" level=error msg="Failed to destroy network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.816194 containerd[1631]: time="2026-04-24T23:34:17.816174948Z" level=error msg="encountered an error cleaning up failed sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.817169 containerd[1631]: time="2026-04-24T23:34:17.817145440Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-kc2pn,Uid:51e78ef8-95fe-4fb0-adbe-918fd6d5713b,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.819116 kubelet[2738]: E0424 23:34:17.817419 2738 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.819116 kubelet[2738]: E0424 23:34:17.817489 2738 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-kc2pn" Apr 24 23:34:17.819116 kubelet[2738]: E0424 23:34:17.817508 2738 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-kc2pn" Apr 24 23:34:17.818930 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29-shm.mount: Deactivated successfully. Apr 24 23:34:17.819294 kubelet[2738]: E0424 23:34:17.817561 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-kc2pn_kube-system(51e78ef8-95fe-4fb0-adbe-918fd6d5713b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-kc2pn_kube-system(51e78ef8-95fe-4fb0-adbe-918fd6d5713b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-kc2pn" podUID="51e78ef8-95fe-4fb0-adbe-918fd6d5713b" Apr 24 23:34:17.856029 containerd[1631]: time="2026-04-24T23:34:17.855981969Z" level=error msg="Failed to destroy network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.857122 containerd[1631]: time="2026-04-24T23:34:17.856543075Z" level=error msg="encountered an error cleaning up failed sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.857238 containerd[1631]: time="2026-04-24T23:34:17.857214579Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cd754fd7f-hzjqr,Uid:5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.857615 kubelet[2738]: E0424 23:34:17.857551 2738 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.857615 kubelet[2738]: E0424 23:34:17.857607 2738 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6cd754fd7f-hzjqr" Apr 24 23:34:17.857824 kubelet[2738]: E0424 23:34:17.857626 2738 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6cd754fd7f-hzjqr" Apr 24 23:34:17.857824 kubelet[2738]: E0424 23:34:17.857665 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6cd754fd7f-hzjqr_calico-system(5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6cd754fd7f-hzjqr_calico-system(5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6cd754fd7f-hzjqr" podUID="5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b" Apr 24 23:34:17.860994 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde-shm.mount: Deactivated successfully. Apr 24 23:34:17.863559 containerd[1631]: time="2026-04-24T23:34:17.863540099Z" level=error msg="Failed to destroy network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.868990 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f-shm.mount: Deactivated successfully. Apr 24 23:34:17.869894 containerd[1631]: time="2026-04-24T23:34:17.869874678Z" level=error msg="encountered an error cleaning up failed sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.870000 containerd[1631]: time="2026-04-24T23:34:17.869985397Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85f77d67c7-7k5mt,Uid:c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.870196 kubelet[2738]: E0424 23:34:17.870176 2738 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.870285 kubelet[2738]: E0424 23:34:17.870274 2738 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-85f77d67c7-7k5mt" Apr 24 23:34:17.870373 kubelet[2738]: E0424 23:34:17.870361 2738 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-85f77d67c7-7k5mt" Apr 24 23:34:17.870480 kubelet[2738]: E0424 23:34:17.870464 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-85f77d67c7-7k5mt_calico-system(c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-85f77d67c7-7k5mt_calico-system(c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-85f77d67c7-7k5mt" podUID="c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b" Apr 24 23:34:17.875803 containerd[1631]: time="2026-04-24T23:34:17.875765501Z" level=error msg="Failed to destroy network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.880180 containerd[1631]: time="2026-04-24T23:34:17.877342578Z" level=error msg="encountered an error cleaning up failed sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.880180 containerd[1631]: time="2026-04-24T23:34:17.877383267Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-54cb6594cd-g4bq8,Uid:fd07f2b5-728d-4d3c-a1de-b9db458623b8,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.880277 kubelet[2738]: E0424 23:34:17.877518 2738 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.880277 kubelet[2738]: E0424 23:34:17.877556 2738 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-54cb6594cd-g4bq8" Apr 24 23:34:17.880277 kubelet[2738]: E0424 23:34:17.877579 2738 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-54cb6594cd-g4bq8" Apr 24 23:34:17.880349 kubelet[2738]: E0424 23:34:17.877626 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-54cb6594cd-g4bq8_calico-system(fd07f2b5-728d-4d3c-a1de-b9db458623b8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-54cb6594cd-g4bq8_calico-system(fd07f2b5-728d-4d3c-a1de-b9db458623b8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-54cb6594cd-g4bq8" podUID="fd07f2b5-728d-4d3c-a1de-b9db458623b8" Apr 24 23:34:17.882530 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a-shm.mount: Deactivated successfully. Apr 24 23:34:17.894856 containerd[1631]: time="2026-04-24T23:34:17.894564630Z" level=error msg="Failed to destroy network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.895071 containerd[1631]: time="2026-04-24T23:34:17.894932417Z" level=error msg="encountered an error cleaning up failed sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.895071 containerd[1631]: time="2026-04-24T23:34:17.894967397Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85f77d67c7-lz87d,Uid:0887e740-8164-4423-b084-a17d88aff0ef,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.895754 kubelet[2738]: E0424 23:34:17.895151 2738 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.895754 kubelet[2738]: E0424 23:34:17.895199 2738 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-85f77d67c7-lz87d" Apr 24 23:34:17.895754 kubelet[2738]: E0424 23:34:17.895216 2738 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-85f77d67c7-lz87d" Apr 24 23:34:17.895902 kubelet[2738]: E0424 23:34:17.895271 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-85f77d67c7-lz87d_calico-system(0887e740-8164-4423-b084-a17d88aff0ef)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-85f77d67c7-lz87d_calico-system(0887e740-8164-4423-b084-a17d88aff0ef)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-85f77d67c7-lz87d" podUID="0887e740-8164-4423-b084-a17d88aff0ef" Apr 24 23:34:17.898502 containerd[1631]: time="2026-04-24T23:34:17.898459959Z" level=error msg="Failed to destroy network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.898751 containerd[1631]: time="2026-04-24T23:34:17.898724877Z" level=error msg="encountered an error cleaning up failed sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.898774 containerd[1631]: time="2026-04-24T23:34:17.898758587Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-l624f,Uid:f187502e-fb21-4086-9792-3a608142290e,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.898993 kubelet[2738]: E0424 23:34:17.898879 2738 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.898993 kubelet[2738]: E0424 23:34:17.898916 2738 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-5b85766d88-l624f" Apr 24 23:34:17.898993 kubelet[2738]: E0424 23:34:17.898928 2738 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-5b85766d88-l624f" Apr 24 23:34:17.899057 kubelet[2738]: E0424 23:34:17.898957 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-5b85766d88-l624f_calico-system(f187502e-fb21-4086-9792-3a608142290e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-5b85766d88-l624f_calico-system(f187502e-fb21-4086-9792-3a608142290e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-5b85766d88-l624f" podUID="f187502e-fb21-4086-9792-3a608142290e" Apr 24 23:34:17.899353 containerd[1631]: time="2026-04-24T23:34:17.899329931Z" level=error msg="Failed to destroy network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.899578 containerd[1631]: time="2026-04-24T23:34:17.899558350Z" level=error msg="encountered an error cleaning up failed sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.899603 containerd[1631]: time="2026-04-24T23:34:17.899585330Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2gd6v,Uid:d8da1862-fd17-4f27-a872-feaa69b7162e,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.899703 kubelet[2738]: E0424 23:34:17.899683 2738 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:17.899740 kubelet[2738]: E0424 23:34:17.899708 2738 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-2gd6v" Apr 24 23:34:17.899740 kubelet[2738]: E0424 23:34:17.899726 2738 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-2gd6v" Apr 24 23:34:17.900176 kubelet[2738]: E0424 23:34:17.899770 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-2gd6v_kube-system(d8da1862-fd17-4f27-a872-feaa69b7162e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-2gd6v_kube-system(d8da1862-fd17-4f27-a872-feaa69b7162e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-2gd6v" podUID="d8da1862-fd17-4f27-a872-feaa69b7162e" Apr 24 23:34:18.122597 containerd[1631]: time="2026-04-24T23:34:18.122346233Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kflpl,Uid:7556f2ac-97a1-4b30-8823-9a8d6f3243fe,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:18.212297 containerd[1631]: time="2026-04-24T23:34:18.212036125Z" level=error msg="Failed to destroy network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.212696 containerd[1631]: time="2026-04-24T23:34:18.212434722Z" level=error msg="encountered an error cleaning up failed sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.212696 containerd[1631]: time="2026-04-24T23:34:18.212473691Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kflpl,Uid:7556f2ac-97a1-4b30-8823-9a8d6f3243fe,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.212830 kubelet[2738]: E0424 23:34:18.212674 2738 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.212830 kubelet[2738]: E0424 23:34:18.212737 2738 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kflpl" Apr 24 23:34:18.212830 kubelet[2738]: E0424 23:34:18.212760 2738 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kflpl" Apr 24 23:34:18.212984 kubelet[2738]: E0424 23:34:18.212807 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kflpl_calico-system(7556f2ac-97a1-4b30-8823-9a8d6f3243fe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kflpl_calico-system(7556f2ac-97a1-4b30-8823-9a8d6f3243fe)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:18.225418 kubelet[2738]: I0424 23:34:18.225381 2738 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:18.226939 containerd[1631]: time="2026-04-24T23:34:18.226556097Z" level=info msg="StopPodSandbox for \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\"" Apr 24 23:34:18.226939 containerd[1631]: time="2026-04-24T23:34:18.226713135Z" level=info msg="Ensure that sandbox 8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb in task-service has been cleanup successfully" Apr 24 23:34:18.229183 kubelet[2738]: I0424 23:34:18.229121 2738 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:18.231242 containerd[1631]: time="2026-04-24T23:34:18.231226362Z" level=info msg="StopPodSandbox for \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\"" Apr 24 23:34:18.231422 containerd[1631]: time="2026-04-24T23:34:18.231410920Z" level=info msg="Ensure that sandbox 84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde in task-service has been cleanup successfully" Apr 24 23:34:18.233752 kubelet[2738]: I0424 23:34:18.231304 2738 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:18.235550 kubelet[2738]: I0424 23:34:18.235332 2738 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:18.235677 containerd[1631]: time="2026-04-24T23:34:18.235663569Z" level=info msg="StopPodSandbox for \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\"" Apr 24 23:34:18.235831 containerd[1631]: time="2026-04-24T23:34:18.235810828Z" level=info msg="Ensure that sandbox 76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29 in task-service has been cleanup successfully" Apr 24 23:34:18.239340 containerd[1631]: time="2026-04-24T23:34:18.239326172Z" level=info msg="StopPodSandbox for \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\"" Apr 24 23:34:18.239608 containerd[1631]: time="2026-04-24T23:34:18.239594750Z" level=info msg="Ensure that sandbox 3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f in task-service has been cleanup successfully" Apr 24 23:34:18.240958 kubelet[2738]: I0424 23:34:18.240946 2738 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:18.241747 containerd[1631]: time="2026-04-24T23:34:18.241402876Z" level=info msg="StopPodSandbox for \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\"" Apr 24 23:34:18.241747 containerd[1631]: time="2026-04-24T23:34:18.241495005Z" level=info msg="Ensure that sandbox 71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20 in task-service has been cleanup successfully" Apr 24 23:34:18.249438 kubelet[2738]: I0424 23:34:18.249405 2738 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:18.251755 containerd[1631]: time="2026-04-24T23:34:18.251734859Z" level=info msg="StopPodSandbox for \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\"" Apr 24 23:34:18.251939 containerd[1631]: time="2026-04-24T23:34:18.251927917Z" level=info msg="Ensure that sandbox 3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f in task-service has been cleanup successfully" Apr 24 23:34:18.255495 kubelet[2738]: I0424 23:34:18.255435 2738 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:18.256175 containerd[1631]: time="2026-04-24T23:34:18.255913177Z" level=info msg="StopPodSandbox for \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\"" Apr 24 23:34:18.256175 containerd[1631]: time="2026-04-24T23:34:18.256013166Z" level=info msg="Ensure that sandbox 41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a in task-service has been cleanup successfully" Apr 24 23:34:18.269668 kubelet[2738]: I0424 23:34:18.269442 2738 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:18.274290 containerd[1631]: time="2026-04-24T23:34:18.274015902Z" level=info msg="StopPodSandbox for \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\"" Apr 24 23:34:18.275836 containerd[1631]: time="2026-04-24T23:34:18.275819299Z" level=info msg="Ensure that sandbox 04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e in task-service has been cleanup successfully" Apr 24 23:34:18.284548 containerd[1631]: time="2026-04-24T23:34:18.284317686Z" level=info msg="CreateContainer within sandbox \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Apr 24 23:34:18.302851 containerd[1631]: time="2026-04-24T23:34:18.302815978Z" level=info msg="CreateContainer within sandbox \"35c4bcdc3df6844f8390f7297a4bb8f33cb5a85cdad21e8c83c620a4343d80dc\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"b5284ea3a90f10dfe43b73c58f5aab2917b4b9dc0fd4594ef13f0a114b391f77\"" Apr 24 23:34:18.304120 containerd[1631]: time="2026-04-24T23:34:18.303300184Z" level=info msg="StartContainer for \"b5284ea3a90f10dfe43b73c58f5aab2917b4b9dc0fd4594ef13f0a114b391f77\"" Apr 24 23:34:18.323777 containerd[1631]: time="2026-04-24T23:34:18.323730782Z" level=error msg="StopPodSandbox for \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\" failed" error="failed to destroy network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.324114 kubelet[2738]: E0424 23:34:18.324066 2738 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:18.324437 kubelet[2738]: E0424 23:34:18.324134 2738 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb"} Apr 24 23:34:18.324437 kubelet[2738]: E0424 23:34:18.324220 2738 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"d8da1862-fd17-4f27-a872-feaa69b7162e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 24 23:34:18.324437 kubelet[2738]: E0424 23:34:18.324243 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"d8da1862-fd17-4f27-a872-feaa69b7162e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-2gd6v" podUID="d8da1862-fd17-4f27-a872-feaa69b7162e" Apr 24 23:34:18.324594 containerd[1631]: time="2026-04-24T23:34:18.324576086Z" level=error msg="StopPodSandbox for \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\" failed" error="failed to destroy network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.324882 kubelet[2738]: E0424 23:34:18.324835 2738 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:18.324926 kubelet[2738]: E0424 23:34:18.324884 2738 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20"} Apr 24 23:34:18.324926 kubelet[2738]: E0424 23:34:18.324901 2738 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"0887e740-8164-4423-b084-a17d88aff0ef\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 24 23:34:18.324926 kubelet[2738]: E0424 23:34:18.324914 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"0887e740-8164-4423-b084-a17d88aff0ef\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-85f77d67c7-lz87d" podUID="0887e740-8164-4423-b084-a17d88aff0ef" Apr 24 23:34:18.346220 containerd[1631]: time="2026-04-24T23:34:18.346183815Z" level=error msg="StopPodSandbox for \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\" failed" error="failed to destroy network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.346607 kubelet[2738]: E0424 23:34:18.346541 2738 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:18.346674 kubelet[2738]: E0424 23:34:18.346621 2738 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a"} Apr 24 23:34:18.346701 kubelet[2738]: E0424 23:34:18.346681 2738 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 24 23:34:18.346757 kubelet[2738]: E0424 23:34:18.346707 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-54cb6594cd-g4bq8" podUID="fd07f2b5-728d-4d3c-a1de-b9db458623b8" Apr 24 23:34:18.351257 containerd[1631]: time="2026-04-24T23:34:18.351183968Z" level=error msg="StopPodSandbox for \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\" failed" error="failed to destroy network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.351763 kubelet[2738]: E0424 23:34:18.351328 2738 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:18.351763 kubelet[2738]: E0424 23:34:18.351369 2738 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde"} Apr 24 23:34:18.351763 kubelet[2738]: E0424 23:34:18.351395 2738 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 24 23:34:18.351763 kubelet[2738]: E0424 23:34:18.351418 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6cd754fd7f-hzjqr" podUID="5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b" Apr 24 23:34:18.365154 containerd[1631]: time="2026-04-24T23:34:18.365089123Z" level=error msg="StopPodSandbox for \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\" failed" error="failed to destroy network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.365606 kubelet[2738]: E0424 23:34:18.365459 2738 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:18.365675 kubelet[2738]: E0424 23:34:18.365656 2738 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29"} Apr 24 23:34:18.365711 kubelet[2738]: E0424 23:34:18.365693 2738 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"51e78ef8-95fe-4fb0-adbe-918fd6d5713b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 24 23:34:18.365880 kubelet[2738]: E0424 23:34:18.365855 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"51e78ef8-95fe-4fb0-adbe-918fd6d5713b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-kc2pn" podUID="51e78ef8-95fe-4fb0-adbe-918fd6d5713b" Apr 24 23:34:18.371138 containerd[1631]: time="2026-04-24T23:34:18.371117428Z" level=error msg="StopPodSandbox for \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\" failed" error="failed to destroy network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.371327 kubelet[2738]: E0424 23:34:18.371300 2738 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:18.371362 kubelet[2738]: E0424 23:34:18.371332 2738 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f"} Apr 24 23:34:18.371362 kubelet[2738]: E0424 23:34:18.371356 2738 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f187502e-fb21-4086-9792-3a608142290e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 24 23:34:18.371362 kubelet[2738]: E0424 23:34:18.371378 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f187502e-fb21-4086-9792-3a608142290e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-5b85766d88-l624f" podUID="f187502e-fb21-4086-9792-3a608142290e" Apr 24 23:34:18.371624 containerd[1631]: time="2026-04-24T23:34:18.371606505Z" level=error msg="StopPodSandbox for \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\" failed" error="failed to destroy network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.372113 kubelet[2738]: E0424 23:34:18.371776 2738 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:18.372113 kubelet[2738]: E0424 23:34:18.371826 2738 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e"} Apr 24 23:34:18.372113 kubelet[2738]: E0424 23:34:18.371849 2738 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"7556f2ac-97a1-4b30-8823-9a8d6f3243fe\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 24 23:34:18.372113 kubelet[2738]: E0424 23:34:18.371870 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"7556f2ac-97a1-4b30-8823-9a8d6f3243fe\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kflpl" podUID="7556f2ac-97a1-4b30-8823-9a8d6f3243fe" Apr 24 23:34:18.373780 containerd[1631]: time="2026-04-24T23:34:18.373762369Z" level=error msg="StopPodSandbox for \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\" failed" error="failed to destroy network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 24 23:34:18.373970 kubelet[2738]: E0424 23:34:18.373943 2738 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:18.374010 kubelet[2738]: E0424 23:34:18.373971 2738 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f"} Apr 24 23:34:18.374036 kubelet[2738]: E0424 23:34:18.374023 2738 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 24 23:34:18.374078 kubelet[2738]: E0424 23:34:18.374047 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-85f77d67c7-7k5mt" podUID="c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b" Apr 24 23:34:18.409424 containerd[1631]: time="2026-04-24T23:34:18.409386773Z" level=info msg="StartContainer for \"b5284ea3a90f10dfe43b73c58f5aab2917b4b9dc0fd4594ef13f0a114b391f77\" returns successfully" Apr 24 23:34:18.768990 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20-shm.mount: Deactivated successfully. Apr 24 23:34:18.769436 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb-shm.mount: Deactivated successfully. Apr 24 23:34:18.769795 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f-shm.mount: Deactivated successfully. Apr 24 23:34:19.282337 containerd[1631]: time="2026-04-24T23:34:19.282255891Z" level=info msg="StopPodSandbox for \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\"" Apr 24 23:34:19.345823 kubelet[2738]: I0424 23:34:19.345494 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-6zx6n" podStartSLOduration=3.164087324 podStartE2EDuration="15.345477283s" podCreationTimestamp="2026-04-24 23:34:04 +0000 UTC" firstStartedPulling="2026-04-24 23:34:04.5621213 +0000 UTC m=+15.534272634" lastFinishedPulling="2026-04-24 23:34:16.743511249 +0000 UTC m=+27.715662593" observedRunningTime="2026-04-24 23:34:19.310362037 +0000 UTC m=+30.282513371" watchObservedRunningTime="2026-04-24 23:34:19.345477283 +0000 UTC m=+30.317628627" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.345 [INFO][4090] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.346 [INFO][4090] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" iface="eth0" netns="/var/run/netns/cni-6627cee7-562a-3d66-c003-2dffaeef3980" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.347 [INFO][4090] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" iface="eth0" netns="/var/run/netns/cni-6627cee7-562a-3d66-c003-2dffaeef3980" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.347 [INFO][4090] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" iface="eth0" netns="/var/run/netns/cni-6627cee7-562a-3d66-c003-2dffaeef3980" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.347 [INFO][4090] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.347 [INFO][4090] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.365 [INFO][4097] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.365 [INFO][4097] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.366 [INFO][4097] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.370 [WARNING][4097] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.370 [INFO][4097] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.371 [INFO][4097] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:19.376310 containerd[1631]: 2026-04-24 23:34:19.374 [INFO][4090] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:19.380097 containerd[1631]: time="2026-04-24T23:34:19.378135206Z" level=info msg="TearDown network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\" successfully" Apr 24 23:34:19.380097 containerd[1631]: time="2026-04-24T23:34:19.378167706Z" level=info msg="StopPodSandbox for \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\" returns successfully" Apr 24 23:34:19.380814 systemd[1]: run-netns-cni\x2d6627cee7\x2d562a\x2d3d66\x2dc003\x2d2dffaeef3980.mount: Deactivated successfully. Apr 24 23:34:19.447283 kubelet[2738]: I0424 23:34:19.447065 2738 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/fd07f2b5-728d-4d3c-a1de-b9db458623b8-nginx-config\") pod \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\" (UID: \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\") " Apr 24 23:34:19.447283 kubelet[2738]: I0424 23:34:19.447149 2738 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/fd07f2b5-728d-4d3c-a1de-b9db458623b8-whisker-backend-key-pair\") pod \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\" (UID: \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\") " Apr 24 23:34:19.447283 kubelet[2738]: I0424 23:34:19.447174 2738 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tw6n\" (UniqueName: \"kubernetes.io/projected/fd07f2b5-728d-4d3c-a1de-b9db458623b8-kube-api-access-4tw6n\") pod \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\" (UID: \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\") " Apr 24 23:34:19.447283 kubelet[2738]: I0424 23:34:19.447193 2738 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd07f2b5-728d-4d3c-a1de-b9db458623b8-whisker-ca-bundle\") pod \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\" (UID: \"fd07f2b5-728d-4d3c-a1de-b9db458623b8\") " Apr 24 23:34:19.447875 kubelet[2738]: I0424 23:34:19.447717 2738 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd07f2b5-728d-4d3c-a1de-b9db458623b8-nginx-config" (OuterVolumeSpecName: "nginx-config") pod "fd07f2b5-728d-4d3c-a1de-b9db458623b8" (UID: "fd07f2b5-728d-4d3c-a1de-b9db458623b8"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 24 23:34:19.447875 kubelet[2738]: I0424 23:34:19.447792 2738 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd07f2b5-728d-4d3c-a1de-b9db458623b8-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "fd07f2b5-728d-4d3c-a1de-b9db458623b8" (UID: "fd07f2b5-728d-4d3c-a1de-b9db458623b8"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 24 23:34:19.452538 kubelet[2738]: I0424 23:34:19.452175 2738 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd07f2b5-728d-4d3c-a1de-b9db458623b8-kube-api-access-4tw6n" (OuterVolumeSpecName: "kube-api-access-4tw6n") pod "fd07f2b5-728d-4d3c-a1de-b9db458623b8" (UID: "fd07f2b5-728d-4d3c-a1de-b9db458623b8"). InnerVolumeSpecName "kube-api-access-4tw6n". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 24 23:34:19.452680 kubelet[2738]: I0424 23:34:19.452664 2738 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd07f2b5-728d-4d3c-a1de-b9db458623b8-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "fd07f2b5-728d-4d3c-a1de-b9db458623b8" (UID: "fd07f2b5-728d-4d3c-a1de-b9db458623b8"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 24 23:34:19.453992 systemd[1]: var-lib-kubelet-pods-fd07f2b5\x2d728d\x2d4d3c\x2da1de\x2db9db458623b8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d4tw6n.mount: Deactivated successfully. Apr 24 23:34:19.454295 systemd[1]: var-lib-kubelet-pods-fd07f2b5\x2d728d\x2d4d3c\x2da1de\x2db9db458623b8-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Apr 24 23:34:19.548385 kubelet[2738]: I0424 23:34:19.548334 2738 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/fd07f2b5-728d-4d3c-a1de-b9db458623b8-nginx-config\") on node \"ci-4081-3-6-n-e88c017743\" DevicePath \"\"" Apr 24 23:34:19.548546 kubelet[2738]: I0424 23:34:19.548393 2738 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/fd07f2b5-728d-4d3c-a1de-b9db458623b8-whisker-backend-key-pair\") on node \"ci-4081-3-6-n-e88c017743\" DevicePath \"\"" Apr 24 23:34:19.548546 kubelet[2738]: I0424 23:34:19.548421 2738 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4tw6n\" (UniqueName: \"kubernetes.io/projected/fd07f2b5-728d-4d3c-a1de-b9db458623b8-kube-api-access-4tw6n\") on node \"ci-4081-3-6-n-e88c017743\" DevicePath \"\"" Apr 24 23:34:19.548546 kubelet[2738]: I0424 23:34:19.548450 2738 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd07f2b5-728d-4d3c-a1de-b9db458623b8-whisker-ca-bundle\") on node \"ci-4081-3-6-n-e88c017743\" DevicePath \"\"" Apr 24 23:34:20.280701 kubelet[2738]: I0424 23:34:20.280654 2738 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 24 23:34:20.453846 kubelet[2738]: I0424 23:34:20.453577 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/ecae16b3-325d-4c58-9c3e-29f85488b55c-nginx-config\") pod \"whisker-7685fb84f5-v5lpn\" (UID: \"ecae16b3-325d-4c58-9c3e-29f85488b55c\") " pod="calico-system/whisker-7685fb84f5-v5lpn" Apr 24 23:34:20.453846 kubelet[2738]: I0424 23:34:20.453638 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ecae16b3-325d-4c58-9c3e-29f85488b55c-whisker-backend-key-pair\") pod \"whisker-7685fb84f5-v5lpn\" (UID: \"ecae16b3-325d-4c58-9c3e-29f85488b55c\") " pod="calico-system/whisker-7685fb84f5-v5lpn" Apr 24 23:34:20.453846 kubelet[2738]: I0424 23:34:20.453681 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecae16b3-325d-4c58-9c3e-29f85488b55c-whisker-ca-bundle\") pod \"whisker-7685fb84f5-v5lpn\" (UID: \"ecae16b3-325d-4c58-9c3e-29f85488b55c\") " pod="calico-system/whisker-7685fb84f5-v5lpn" Apr 24 23:34:20.453846 kubelet[2738]: I0424 23:34:20.453709 2738 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26khk\" (UniqueName: \"kubernetes.io/projected/ecae16b3-325d-4c58-9c3e-29f85488b55c-kube-api-access-26khk\") pod \"whisker-7685fb84f5-v5lpn\" (UID: \"ecae16b3-325d-4c58-9c3e-29f85488b55c\") " pod="calico-system/whisker-7685fb84f5-v5lpn" Apr 24 23:34:20.664347 containerd[1631]: time="2026-04-24T23:34:20.664266946Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7685fb84f5-v5lpn,Uid:ecae16b3-325d-4c58-9c3e-29f85488b55c,Namespace:calico-system,Attempt:0,}" Apr 24 23:34:20.807237 systemd-networkd[1262]: calif64de6f180a: Link UP Apr 24 23:34:20.809759 systemd-networkd[1262]: calif64de6f180a: Gained carrier Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.718 [ERROR][4202] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.728 [INFO][4202] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0 whisker-7685fb84f5- calico-system ecae16b3-325d-4c58-9c3e-29f85488b55c 918 0 2026-04-24 23:34:20 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:7685fb84f5 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4081-3-6-n-e88c017743 whisker-7685fb84f5-v5lpn eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calif64de6f180a [] [] }} ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Namespace="calico-system" Pod="whisker-7685fb84f5-v5lpn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.728 [INFO][4202] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Namespace="calico-system" Pod="whisker-7685fb84f5-v5lpn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.758 [INFO][4216] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" HandleID="k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.766 [INFO][4216] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" HandleID="k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002777d0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-e88c017743", "pod":"whisker-7685fb84f5-v5lpn", "timestamp":"2026-04-24 23:34:20.758784085 +0000 UTC"}, Hostname:"ci-4081-3-6-n-e88c017743", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00030b760)} Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.766 [INFO][4216] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.766 [INFO][4216] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.766 [INFO][4216] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-e88c017743' Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.769 [INFO][4216] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.773 [INFO][4216] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.779 [INFO][4216] ipam/ipam.go 526: Trying affinity for 192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.780 [INFO][4216] ipam/ipam.go 160: Attempting to load block cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.781 [INFO][4216] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.781 [INFO][4216] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.115.0/26 handle="k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.783 [INFO][4216] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9 Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.787 [INFO][4216] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.115.0/26 handle="k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.791 [INFO][4216] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.115.1/26] block=192.168.115.0/26 handle="k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.791 [INFO][4216] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.115.1/26] handle="k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.791 [INFO][4216] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:20.830182 containerd[1631]: 2026-04-24 23:34:20.791 [INFO][4216] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.115.1/26] IPv6=[] ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" HandleID="k8s-pod-network.0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" Apr 24 23:34:20.830664 containerd[1631]: 2026-04-24 23:34:20.794 [INFO][4202] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Namespace="calico-system" Pod="whisker-7685fb84f5-v5lpn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0", GenerateName:"whisker-7685fb84f5-", Namespace:"calico-system", SelfLink:"", UID:"ecae16b3-325d-4c58-9c3e-29f85488b55c", ResourceVersion:"918", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7685fb84f5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"", Pod:"whisker-7685fb84f5-v5lpn", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.115.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calif64de6f180a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:20.830664 containerd[1631]: 2026-04-24 23:34:20.794 [INFO][4202] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.115.1/32] ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Namespace="calico-system" Pod="whisker-7685fb84f5-v5lpn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" Apr 24 23:34:20.830664 containerd[1631]: 2026-04-24 23:34:20.794 [INFO][4202] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif64de6f180a ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Namespace="calico-system" Pod="whisker-7685fb84f5-v5lpn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" Apr 24 23:34:20.830664 containerd[1631]: 2026-04-24 23:34:20.811 [INFO][4202] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Namespace="calico-system" Pod="whisker-7685fb84f5-v5lpn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" Apr 24 23:34:20.830664 containerd[1631]: 2026-04-24 23:34:20.811 [INFO][4202] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Namespace="calico-system" Pod="whisker-7685fb84f5-v5lpn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0", GenerateName:"whisker-7685fb84f5-", Namespace:"calico-system", SelfLink:"", UID:"ecae16b3-325d-4c58-9c3e-29f85488b55c", ResourceVersion:"918", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7685fb84f5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9", Pod:"whisker-7685fb84f5-v5lpn", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.115.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calif64de6f180a", MAC:"62:f0:b7:8b:d3:8e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:20.830664 containerd[1631]: 2026-04-24 23:34:20.821 [INFO][4202] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9" Namespace="calico-system" Pod="whisker-7685fb84f5-v5lpn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--7685fb84f5--v5lpn-eth0" Apr 24 23:34:20.849160 containerd[1631]: time="2026-04-24T23:34:20.849067413Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:20.849262 containerd[1631]: time="2026-04-24T23:34:20.849158353Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:20.849262 containerd[1631]: time="2026-04-24T23:34:20.849191052Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:20.849364 containerd[1631]: time="2026-04-24T23:34:20.849340371Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:20.929073 containerd[1631]: time="2026-04-24T23:34:20.928983097Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7685fb84f5-v5lpn,Uid:ecae16b3-325d-4c58-9c3e-29f85488b55c,Namespace:calico-system,Attempt:0,} returns sandbox id \"0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9\"" Apr 24 23:34:20.932995 containerd[1631]: time="2026-04-24T23:34:20.932971962Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\"" Apr 24 23:34:21.121271 kubelet[2738]: I0424 23:34:21.121119 2738 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd07f2b5-728d-4d3c-a1de-b9db458623b8" path="/var/lib/kubelet/pods/fd07f2b5-728d-4d3c-a1de-b9db458623b8/volumes" Apr 24 23:34:21.884036 kubelet[2738]: I0424 23:34:21.883931 2738 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 24 23:34:22.094132 kernel: calico-node[4312]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Apr 24 23:34:22.352218 systemd-networkd[1262]: calif64de6f180a: Gained IPv6LL Apr 24 23:34:22.480226 systemd-networkd[1262]: vxlan.calico: Link UP Apr 24 23:34:22.480233 systemd-networkd[1262]: vxlan.calico: Gained carrier Apr 24 23:34:22.613212 containerd[1631]: time="2026-04-24T23:34:22.612355141Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:22.614221 containerd[1631]: time="2026-04-24T23:34:22.614187021Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.4: active requests=0, bytes read=6039889" Apr 24 23:34:22.615679 containerd[1631]: time="2026-04-24T23:34:22.615661852Z" level=info msg="ImageCreate event name:\"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:22.617850 containerd[1631]: time="2026-04-24T23:34:22.617832910Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:22.618874 containerd[1631]: time="2026-04-24T23:34:22.618857225Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.4\" with image id \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\", size \"7595926\" in 1.685857753s" Apr 24 23:34:22.618957 containerd[1631]: time="2026-04-24T23:34:22.618946004Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\" returns image reference \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\"" Apr 24 23:34:22.624205 containerd[1631]: time="2026-04-24T23:34:22.623981186Z" level=info msg="CreateContainer within sandbox \"0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Apr 24 23:34:22.637436 containerd[1631]: time="2026-04-24T23:34:22.637405832Z" level=info msg="CreateContainer within sandbox \"0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"2de94891d289869cd91b4ef40a2971a27f06431b98c7e4f8f466106b006a526b\"" Apr 24 23:34:22.637966 containerd[1631]: time="2026-04-24T23:34:22.637945328Z" level=info msg="StartContainer for \"2de94891d289869cd91b4ef40a2971a27f06431b98c7e4f8f466106b006a526b\"" Apr 24 23:34:22.673184 systemd[1]: run-containerd-runc-k8s.io-2de94891d289869cd91b4ef40a2971a27f06431b98c7e4f8f466106b006a526b-runc.xHEmL4.mount: Deactivated successfully. Apr 24 23:34:22.729848 containerd[1631]: time="2026-04-24T23:34:22.729795827Z" level=info msg="StartContainer for \"2de94891d289869cd91b4ef40a2971a27f06431b98c7e4f8f466106b006a526b\" returns successfully" Apr 24 23:34:22.734077 containerd[1631]: time="2026-04-24T23:34:22.734047783Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\"" Apr 24 23:34:24.209292 systemd-networkd[1262]: vxlan.calico: Gained IPv6LL Apr 24 23:34:24.565503 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1380418308.mount: Deactivated successfully. Apr 24 23:34:24.580354 containerd[1631]: time="2026-04-24T23:34:24.580316421Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:24.581465 containerd[1631]: time="2026-04-24T23:34:24.581380926Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.4: active requests=0, bytes read=17609475" Apr 24 23:34:24.582479 containerd[1631]: time="2026-04-24T23:34:24.582306732Z" level=info msg="ImageCreate event name:\"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:24.584220 containerd[1631]: time="2026-04-24T23:34:24.584201252Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:24.584686 containerd[1631]: time="2026-04-24T23:34:24.584663791Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" with image id \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\", size \"17609305\" in 1.850590018s" Apr 24 23:34:24.584717 containerd[1631]: time="2026-04-24T23:34:24.584690540Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" returns image reference \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\"" Apr 24 23:34:24.587587 containerd[1631]: time="2026-04-24T23:34:24.587560056Z" level=info msg="CreateContainer within sandbox \"0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Apr 24 23:34:24.601640 containerd[1631]: time="2026-04-24T23:34:24.601610829Z" level=info msg="CreateContainer within sandbox \"0b8b27c41c2054106df38d1525370dda86385b79979ce86b2fb1aea223fe29b9\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"7879daed52250af92a8a089fffa6db4c0371b7f8cdec404b39cbe03fe3c3bcf1\"" Apr 24 23:34:24.602018 containerd[1631]: time="2026-04-24T23:34:24.601945617Z" level=info msg="StartContainer for \"7879daed52250af92a8a089fffa6db4c0371b7f8cdec404b39cbe03fe3c3bcf1\"" Apr 24 23:34:24.660644 containerd[1631]: time="2026-04-24T23:34:24.660609766Z" level=info msg="StartContainer for \"7879daed52250af92a8a089fffa6db4c0371b7f8cdec404b39cbe03fe3c3bcf1\" returns successfully" Apr 24 23:34:25.311113 kubelet[2738]: I0424 23:34:25.310830 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-7685fb84f5-v5lpn" podStartSLOduration=1.656317198 podStartE2EDuration="5.31078853s" podCreationTimestamp="2026-04-24 23:34:20 +0000 UTC" firstStartedPulling="2026-04-24 23:34:20.931037684 +0000 UTC m=+31.903189028" lastFinishedPulling="2026-04-24 23:34:24.585509026 +0000 UTC m=+35.557660360" observedRunningTime="2026-04-24 23:34:25.308726159 +0000 UTC m=+36.280877533" watchObservedRunningTime="2026-04-24 23:34:25.31078853 +0000 UTC m=+36.282939874" Apr 24 23:34:25.410828 systemd[1]: run-containerd-runc-k8s.io-7879daed52250af92a8a089fffa6db4c0371b7f8cdec404b39cbe03fe3c3bcf1-runc.bf4IrO.mount: Deactivated successfully. Apr 24 23:34:29.120070 containerd[1631]: time="2026-04-24T23:34:29.119261267Z" level=info msg="StopPodSandbox for \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\"" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.188 [INFO][4546] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.188 [INFO][4546] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" iface="eth0" netns="/var/run/netns/cni-39d6b144-f739-9a83-5df3-5087b4b561a9" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.190 [INFO][4546] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" iface="eth0" netns="/var/run/netns/cni-39d6b144-f739-9a83-5df3-5087b4b561a9" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.191 [INFO][4546] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" iface="eth0" netns="/var/run/netns/cni-39d6b144-f739-9a83-5df3-5087b4b561a9" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.191 [INFO][4546] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.191 [INFO][4546] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.209 [INFO][4554] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.209 [INFO][4554] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.209 [INFO][4554] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.214 [WARNING][4554] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.214 [INFO][4554] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.216 [INFO][4554] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:29.220770 containerd[1631]: 2026-04-24 23:34:29.218 [INFO][4546] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:29.222282 containerd[1631]: time="2026-04-24T23:34:29.222147743Z" level=info msg="TearDown network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\" successfully" Apr 24 23:34:29.222282 containerd[1631]: time="2026-04-24T23:34:29.222171643Z" level=info msg="StopPodSandbox for \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\" returns successfully" Apr 24 23:34:29.223580 containerd[1631]: time="2026-04-24T23:34:29.223556329Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85f77d67c7-lz87d,Uid:0887e740-8164-4423-b084-a17d88aff0ef,Namespace:calico-system,Attempt:1,}" Apr 24 23:34:29.224843 systemd[1]: run-netns-cni\x2d39d6b144\x2df739\x2d9a83\x2d5df3\x2d5087b4b561a9.mount: Deactivated successfully. Apr 24 23:34:29.332017 systemd-networkd[1262]: cali9b95f792bb7: Link UP Apr 24 23:34:29.332250 systemd-networkd[1262]: cali9b95f792bb7: Gained carrier Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.279 [INFO][4560] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0 calico-apiserver-85f77d67c7- calico-system 0887e740-8164-4423-b084-a17d88aff0ef 963 0 2026-04-24 23:34:03 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:85f77d67c7 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081-3-6-n-e88c017743 calico-apiserver-85f77d67c7-lz87d eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali9b95f792bb7 [] [] }} ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-lz87d" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.279 [INFO][4560] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-lz87d" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.298 [INFO][4572] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" HandleID="k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.302 [INFO][4572] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" HandleID="k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000277440), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-e88c017743", "pod":"calico-apiserver-85f77d67c7-lz87d", "timestamp":"2026-04-24 23:34:29.298115987 +0000 UTC"}, Hostname:"ci-4081-3-6-n-e88c017743", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00022cf20)} Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.302 [INFO][4572] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.302 [INFO][4572] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.302 [INFO][4572] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-e88c017743' Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.304 [INFO][4572] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.307 [INFO][4572] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.310 [INFO][4572] ipam/ipam.go 526: Trying affinity for 192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.312 [INFO][4572] ipam/ipam.go 160: Attempting to load block cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.313 [INFO][4572] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.314 [INFO][4572] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.115.0/26 handle="k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.315 [INFO][4572] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82 Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.317 [INFO][4572] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.115.0/26 handle="k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.321 [INFO][4572] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.115.2/26] block=192.168.115.0/26 handle="k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.321 [INFO][4572] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.115.2/26] handle="k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.321 [INFO][4572] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:29.340682 containerd[1631]: 2026-04-24 23:34:29.321 [INFO][4572] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.115.2/26] IPv6=[] ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" HandleID="k8s-pod-network.d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.341734 containerd[1631]: 2026-04-24 23:34:29.325 [INFO][4560] cni-plugin/k8s.go 418: Populated endpoint ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-lz87d" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0", GenerateName:"calico-apiserver-85f77d67c7-", Namespace:"calico-system", SelfLink:"", UID:"0887e740-8164-4423-b084-a17d88aff0ef", ResourceVersion:"963", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85f77d67c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"", Pod:"calico-apiserver-85f77d67c7-lz87d", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.115.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali9b95f792bb7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:29.341734 containerd[1631]: 2026-04-24 23:34:29.325 [INFO][4560] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.115.2/32] ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-lz87d" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.341734 containerd[1631]: 2026-04-24 23:34:29.325 [INFO][4560] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9b95f792bb7 ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-lz87d" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.341734 containerd[1631]: 2026-04-24 23:34:29.329 [INFO][4560] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-lz87d" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.341734 containerd[1631]: 2026-04-24 23:34:29.330 [INFO][4560] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-lz87d" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0", GenerateName:"calico-apiserver-85f77d67c7-", Namespace:"calico-system", SelfLink:"", UID:"0887e740-8164-4423-b084-a17d88aff0ef", ResourceVersion:"963", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85f77d67c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82", Pod:"calico-apiserver-85f77d67c7-lz87d", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.115.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali9b95f792bb7", MAC:"de:38:6b:81:5f:b2", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:29.341734 containerd[1631]: 2026-04-24 23:34:29.338 [INFO][4560] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-lz87d" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:29.382709 containerd[1631]: time="2026-04-24T23:34:29.382517184Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:29.383321 containerd[1631]: time="2026-04-24T23:34:29.383191771Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:29.383321 containerd[1631]: time="2026-04-24T23:34:29.383220371Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:29.384117 containerd[1631]: time="2026-04-24T23:34:29.383434360Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:29.426519 kubelet[2738]: I0424 23:34:29.426432 2738 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 24 23:34:29.435637 containerd[1631]: time="2026-04-24T23:34:29.435602822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85f77d67c7-lz87d,Uid:0887e740-8164-4423-b084-a17d88aff0ef,Namespace:calico-system,Attempt:1,} returns sandbox id \"d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82\"" Apr 24 23:34:29.439737 containerd[1631]: time="2026-04-24T23:34:29.439250770Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Apr 24 23:34:30.118625 containerd[1631]: time="2026-04-24T23:34:30.118212089Z" level=info msg="StopPodSandbox for \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\"" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.186 [INFO][4694] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.187 [INFO][4694] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" iface="eth0" netns="/var/run/netns/cni-8393a8b8-c1ea-4fc0-905e-a446e859be7c" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.188 [INFO][4694] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" iface="eth0" netns="/var/run/netns/cni-8393a8b8-c1ea-4fc0-905e-a446e859be7c" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.189 [INFO][4694] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" iface="eth0" netns="/var/run/netns/cni-8393a8b8-c1ea-4fc0-905e-a446e859be7c" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.189 [INFO][4694] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.189 [INFO][4694] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.219 [INFO][4701] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.219 [INFO][4701] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.219 [INFO][4701] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.226 [WARNING][4701] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.226 [INFO][4701] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.227 [INFO][4701] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:30.231348 containerd[1631]: 2026-04-24 23:34:30.229 [INFO][4694] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:30.234117 containerd[1631]: time="2026-04-24T23:34:30.233163255Z" level=info msg="TearDown network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\" successfully" Apr 24 23:34:30.234117 containerd[1631]: time="2026-04-24T23:34:30.233196695Z" level=info msg="StopPodSandbox for \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\" returns successfully" Apr 24 23:34:30.234117 containerd[1631]: time="2026-04-24T23:34:30.233694514Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2gd6v,Uid:d8da1862-fd17-4f27-a872-feaa69b7162e,Namespace:kube-system,Attempt:1,}" Apr 24 23:34:30.234022 systemd[1]: run-netns-cni\x2d8393a8b8\x2dc1ea\x2d4fc0\x2d905e\x2da446e859be7c.mount: Deactivated successfully. Apr 24 23:34:30.330116 systemd-networkd[1262]: cali02ca9128171: Link UP Apr 24 23:34:30.330348 systemd-networkd[1262]: cali02ca9128171: Gained carrier Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.270 [INFO][4708] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0 coredns-674b8bbfcf- kube-system d8da1862-fd17-4f27-a872-feaa69b7162e 972 0 2026-04-24 23:33:54 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081-3-6-n-e88c017743 coredns-674b8bbfcf-2gd6v eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali02ca9128171 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Namespace="kube-system" Pod="coredns-674b8bbfcf-2gd6v" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.270 [INFO][4708] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Namespace="kube-system" Pod="coredns-674b8bbfcf-2gd6v" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.293 [INFO][4719] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" HandleID="k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.298 [INFO][4719] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" HandleID="k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ef830), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081-3-6-n-e88c017743", "pod":"coredns-674b8bbfcf-2gd6v", "timestamp":"2026-04-24 23:34:30.293546855 +0000 UTC"}, Hostname:"ci-4081-3-6-n-e88c017743", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000534f20)} Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.298 [INFO][4719] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.298 [INFO][4719] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.298 [INFO][4719] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-e88c017743' Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.300 [INFO][4719] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.303 [INFO][4719] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.307 [INFO][4719] ipam/ipam.go 526: Trying affinity for 192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.308 [INFO][4719] ipam/ipam.go 160: Attempting to load block cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.311 [INFO][4719] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.311 [INFO][4719] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.115.0/26 handle="k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.313 [INFO][4719] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9 Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.316 [INFO][4719] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.115.0/26 handle="k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.322 [INFO][4719] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.115.3/26] block=192.168.115.0/26 handle="k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.322 [INFO][4719] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.115.3/26] handle="k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.322 [INFO][4719] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:30.345297 containerd[1631]: 2026-04-24 23:34:30.322 [INFO][4719] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.115.3/26] IPv6=[] ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" HandleID="k8s-pod-network.633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.345894 containerd[1631]: 2026-04-24 23:34:30.325 [INFO][4708] cni-plugin/k8s.go 418: Populated endpoint ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Namespace="kube-system" Pod="coredns-674b8bbfcf-2gd6v" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"d8da1862-fd17-4f27-a872-feaa69b7162e", ResourceVersion:"972", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 33, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"", Pod:"coredns-674b8bbfcf-2gd6v", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.115.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali02ca9128171", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:30.345894 containerd[1631]: 2026-04-24 23:34:30.325 [INFO][4708] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.115.3/32] ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Namespace="kube-system" Pod="coredns-674b8bbfcf-2gd6v" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.345894 containerd[1631]: 2026-04-24 23:34:30.325 [INFO][4708] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali02ca9128171 ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Namespace="kube-system" Pod="coredns-674b8bbfcf-2gd6v" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.345894 containerd[1631]: 2026-04-24 23:34:30.328 [INFO][4708] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Namespace="kube-system" Pod="coredns-674b8bbfcf-2gd6v" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.345894 containerd[1631]: 2026-04-24 23:34:30.328 [INFO][4708] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Namespace="kube-system" Pod="coredns-674b8bbfcf-2gd6v" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"d8da1862-fd17-4f27-a872-feaa69b7162e", ResourceVersion:"972", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 33, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9", Pod:"coredns-674b8bbfcf-2gd6v", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.115.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali02ca9128171", MAC:"be:5c:15:85:d1:8d", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:30.345894 containerd[1631]: 2026-04-24 23:34:30.338 [INFO][4708] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9" Namespace="kube-system" Pod="coredns-674b8bbfcf-2gd6v" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:30.353015 systemd-networkd[1262]: cali9b95f792bb7: Gained IPv6LL Apr 24 23:34:30.376370 containerd[1631]: time="2026-04-24T23:34:30.375640539Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:30.376370 containerd[1631]: time="2026-04-24T23:34:30.375679679Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:30.376370 containerd[1631]: time="2026-04-24T23:34:30.375689979Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:30.376370 containerd[1631]: time="2026-04-24T23:34:30.375757649Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:30.460845 containerd[1631]: time="2026-04-24T23:34:30.460749125Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2gd6v,Uid:d8da1862-fd17-4f27-a872-feaa69b7162e,Namespace:kube-system,Attempt:1,} returns sandbox id \"633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9\"" Apr 24 23:34:30.464567 containerd[1631]: time="2026-04-24T23:34:30.464488014Z" level=info msg="CreateContainer within sandbox \"633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 24 23:34:30.473513 containerd[1631]: time="2026-04-24T23:34:30.472827129Z" level=info msg="CreateContainer within sandbox \"633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"366868aff06500fd8758a3ce990846c28b1033e5e4e1a28621fa00b708966768\"" Apr 24 23:34:30.473513 containerd[1631]: time="2026-04-24T23:34:30.473215087Z" level=info msg="StartContainer for \"366868aff06500fd8758a3ce990846c28b1033e5e4e1a28621fa00b708966768\"" Apr 24 23:34:30.512940 containerd[1631]: time="2026-04-24T23:34:30.512911799Z" level=info msg="StartContainer for \"366868aff06500fd8758a3ce990846c28b1033e5e4e1a28621fa00b708966768\" returns successfully" Apr 24 23:34:31.135738 containerd[1631]: time="2026-04-24T23:34:31.135698541Z" level=info msg="StopPodSandbox for \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\"" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.200 [INFO][4845] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.200 [INFO][4845] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" iface="eth0" netns="/var/run/netns/cni-360b8131-0fe6-62a9-13ca-896883056772" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.201 [INFO][4845] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" iface="eth0" netns="/var/run/netns/cni-360b8131-0fe6-62a9-13ca-896883056772" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.202 [INFO][4845] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" iface="eth0" netns="/var/run/netns/cni-360b8131-0fe6-62a9-13ca-896883056772" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.202 [INFO][4845] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.202 [INFO][4845] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.237 [INFO][4856] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.240 [INFO][4856] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.240 [INFO][4856] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.247 [WARNING][4856] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.247 [INFO][4856] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.249 [INFO][4856] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:31.257231 containerd[1631]: 2026-04-24 23:34:31.253 [INFO][4845] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:31.260015 systemd[1]: run-netns-cni\x2d360b8131\x2d0fe6\x2d62a9\x2d13ca\x2d896883056772.mount: Deactivated successfully. Apr 24 23:34:31.260481 containerd[1631]: time="2026-04-24T23:34:31.260141589Z" level=info msg="TearDown network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\" successfully" Apr 24 23:34:31.260481 containerd[1631]: time="2026-04-24T23:34:31.260170759Z" level=info msg="StopPodSandbox for \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\" returns successfully" Apr 24 23:34:31.261505 containerd[1631]: time="2026-04-24T23:34:31.261450075Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kflpl,Uid:7556f2ac-97a1-4b30-8823-9a8d6f3243fe,Namespace:calico-system,Attempt:1,}" Apr 24 23:34:31.336929 kubelet[2738]: I0424 23:34:31.336879 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-2gd6v" podStartSLOduration=37.336865928 podStartE2EDuration="37.336865928s" podCreationTimestamp="2026-04-24 23:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-24 23:34:31.336228519 +0000 UTC m=+42.308379853" watchObservedRunningTime="2026-04-24 23:34:31.336865928 +0000 UTC m=+42.309017272" Apr 24 23:34:31.418062 systemd-networkd[1262]: calief231f26ce3: Link UP Apr 24 23:34:31.423287 systemd-networkd[1262]: calief231f26ce3: Gained carrier Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.314 [INFO][4862] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0 csi-node-driver- calico-system 7556f2ac-97a1-4b30-8823-9a8d6f3243fe 983 0 2026-04-24 23:34:04 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6d9d697c7c k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081-3-6-n-e88c017743 csi-node-driver-kflpl eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calief231f26ce3 [] [] }} ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Namespace="calico-system" Pod="csi-node-driver-kflpl" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.314 [INFO][4862] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Namespace="calico-system" Pod="csi-node-driver-kflpl" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.364 [INFO][4874] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" HandleID="k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.374 [INFO][4874] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" HandleID="k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fba0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-e88c017743", "pod":"csi-node-driver-kflpl", "timestamp":"2026-04-24 23:34:31.364156883 +0000 UTC"}, Hostname:"ci-4081-3-6-n-e88c017743", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00030d760)} Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.375 [INFO][4874] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.375 [INFO][4874] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.375 [INFO][4874] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-e88c017743' Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.379 [INFO][4874] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.384 [INFO][4874] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.387 [INFO][4874] ipam/ipam.go 526: Trying affinity for 192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.389 [INFO][4874] ipam/ipam.go 160: Attempting to load block cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.390 [INFO][4874] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.391 [INFO][4874] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.115.0/26 handle="k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.392 [INFO][4874] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.395 [INFO][4874] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.115.0/26 handle="k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.405 [INFO][4874] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.115.4/26] block=192.168.115.0/26 handle="k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.405 [INFO][4874] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.115.4/26] handle="k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.405 [INFO][4874] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:31.448531 containerd[1631]: 2026-04-24 23:34:31.405 [INFO][4874] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.115.4/26] IPv6=[] ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" HandleID="k8s-pod-network.a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.449488 containerd[1631]: 2026-04-24 23:34:31.410 [INFO][4862] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Namespace="calico-system" Pod="csi-node-driver-kflpl" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7556f2ac-97a1-4b30-8823-9a8d6f3243fe", ResourceVersion:"983", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"", Pod:"csi-node-driver-kflpl", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.115.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calief231f26ce3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:31.449488 containerd[1631]: 2026-04-24 23:34:31.410 [INFO][4862] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.115.4/32] ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Namespace="calico-system" Pod="csi-node-driver-kflpl" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.449488 containerd[1631]: 2026-04-24 23:34:31.410 [INFO][4862] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calief231f26ce3 ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Namespace="calico-system" Pod="csi-node-driver-kflpl" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.449488 containerd[1631]: 2026-04-24 23:34:31.423 [INFO][4862] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Namespace="calico-system" Pod="csi-node-driver-kflpl" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.449488 containerd[1631]: 2026-04-24 23:34:31.426 [INFO][4862] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Namespace="calico-system" Pod="csi-node-driver-kflpl" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7556f2ac-97a1-4b30-8823-9a8d6f3243fe", ResourceVersion:"983", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b", Pod:"csi-node-driver-kflpl", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.115.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calief231f26ce3", MAC:"de:86:a5:45:d7:a9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:31.449488 containerd[1631]: 2026-04-24 23:34:31.441 [INFO][4862] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b" Namespace="calico-system" Pod="csi-node-driver-kflpl" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:31.491638 containerd[1631]: time="2026-04-24T23:34:31.491524533Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:31.491638 containerd[1631]: time="2026-04-24T23:34:31.491577183Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:31.491638 containerd[1631]: time="2026-04-24T23:34:31.491588113Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:31.492001 containerd[1631]: time="2026-04-24T23:34:31.491678872Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:31.543137 containerd[1631]: time="2026-04-24T23:34:31.542838931Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kflpl,Uid:7556f2ac-97a1-4b30-8823-9a8d6f3243fe,Namespace:calico-system,Attempt:1,} returns sandbox id \"a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b\"" Apr 24 23:34:31.952239 systemd-networkd[1262]: cali02ca9128171: Gained IPv6LL Apr 24 23:34:31.979605 containerd[1631]: time="2026-04-24T23:34:31.977245807Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:31.986561 containerd[1631]: time="2026-04-24T23:34:31.986528522Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=48415780" Apr 24 23:34:31.997460 containerd[1631]: time="2026-04-24T23:34:31.995675327Z" level=info msg="ImageCreate event name:\"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:32.009970 containerd[1631]: time="2026-04-24T23:34:32.009492081Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:32.010904 containerd[1631]: time="2026-04-24T23:34:32.010167950Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"49971841\" in 2.57089456s" Apr 24 23:34:32.010904 containerd[1631]: time="2026-04-24T23:34:32.010194539Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\"" Apr 24 23:34:32.012154 containerd[1631]: time="2026-04-24T23:34:32.011588606Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\"" Apr 24 23:34:32.013846 containerd[1631]: time="2026-04-24T23:34:32.013816240Z" level=info msg="CreateContainer within sandbox \"d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 24 23:34:32.034488 containerd[1631]: time="2026-04-24T23:34:32.034438298Z" level=info msg="CreateContainer within sandbox \"d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"b36fb88fcd9a6a8d330756550cee73f2b986cbd976ffb86a3ebeae37ec1a0bed\"" Apr 24 23:34:32.035140 containerd[1631]: time="2026-04-24T23:34:32.035018036Z" level=info msg="StartContainer for \"b36fb88fcd9a6a8d330756550cee73f2b986cbd976ffb86a3ebeae37ec1a0bed\"" Apr 24 23:34:32.094280 containerd[1631]: time="2026-04-24T23:34:32.094238857Z" level=info msg="StartContainer for \"b36fb88fcd9a6a8d330756550cee73f2b986cbd976ffb86a3ebeae37ec1a0bed\" returns successfully" Apr 24 23:34:32.118652 containerd[1631]: time="2026-04-24T23:34:32.118300006Z" level=info msg="StopPodSandbox for \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\"" Apr 24 23:34:32.118652 containerd[1631]: time="2026-04-24T23:34:32.118380946Z" level=info msg="StopPodSandbox for \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\"" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.192 [INFO][5008] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.192 [INFO][5008] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" iface="eth0" netns="/var/run/netns/cni-3c736865-6b2f-dcc3-31a1-28483fc3d6b6" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.192 [INFO][5008] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" iface="eth0" netns="/var/run/netns/cni-3c736865-6b2f-dcc3-31a1-28483fc3d6b6" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.192 [INFO][5008] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" iface="eth0" netns="/var/run/netns/cni-3c736865-6b2f-dcc3-31a1-28483fc3d6b6" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.192 [INFO][5008] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.192 [INFO][5008] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.214 [INFO][5028] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.215 [INFO][5028] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.215 [INFO][5028] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.226 [WARNING][5028] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.226 [INFO][5028] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.230 [INFO][5028] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:32.235177 containerd[1631]: 2026-04-24 23:34:32.232 [INFO][5008] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:32.235177 containerd[1631]: time="2026-04-24T23:34:32.234661163Z" level=info msg="TearDown network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\" successfully" Apr 24 23:34:32.235177 containerd[1631]: time="2026-04-24T23:34:32.234690673Z" level=info msg="StopPodSandbox for \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\" returns successfully" Apr 24 23:34:32.239738 containerd[1631]: time="2026-04-24T23:34:32.239470621Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cd754fd7f-hzjqr,Uid:5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b,Namespace:calico-system,Attempt:1,}" Apr 24 23:34:32.241557 systemd[1]: run-netns-cni\x2d3c736865\x2d6b2f\x2ddcc3\x2d31a1\x2d28483fc3d6b6.mount: Deactivated successfully. Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.189 [INFO][5009] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.189 [INFO][5009] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" iface="eth0" netns="/var/run/netns/cni-d491eca0-824b-03a2-a116-bf7c276cb824" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.190 [INFO][5009] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" iface="eth0" netns="/var/run/netns/cni-d491eca0-824b-03a2-a116-bf7c276cb824" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.190 [INFO][5009] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" iface="eth0" netns="/var/run/netns/cni-d491eca0-824b-03a2-a116-bf7c276cb824" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.190 [INFO][5009] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.190 [INFO][5009] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.214 [INFO][5026] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.215 [INFO][5026] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.228 [INFO][5026] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.241 [WARNING][5026] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.241 [INFO][5026] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.243 [INFO][5026] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:32.255558 containerd[1631]: 2026-04-24 23:34:32.247 [INFO][5009] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:32.257122 containerd[1631]: time="2026-04-24T23:34:32.256121659Z" level=info msg="TearDown network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\" successfully" Apr 24 23:34:32.257122 containerd[1631]: time="2026-04-24T23:34:32.256139089Z" level=info msg="StopPodSandbox for \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\" returns successfully" Apr 24 23:34:32.257585 containerd[1631]: time="2026-04-24T23:34:32.257321026Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-kc2pn,Uid:51e78ef8-95fe-4fb0-adbe-918fd6d5713b,Namespace:kube-system,Attempt:1,}" Apr 24 23:34:32.264066 systemd[1]: run-netns-cni\x2dd491eca0\x2d824b\x2d03a2\x2da116\x2dbf7c276cb824.mount: Deactivated successfully. Apr 24 23:34:32.431992 systemd-networkd[1262]: cali92655be1325: Link UP Apr 24 23:34:32.436535 systemd-networkd[1262]: cali92655be1325: Gained carrier Apr 24 23:34:32.439247 kubelet[2738]: I0424 23:34:32.437872 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-85f77d67c7-lz87d" podStartSLOduration=26.863922268 podStartE2EDuration="29.437843529s" podCreationTimestamp="2026-04-24 23:34:03 +0000 UTC" firstStartedPulling="2026-04-24 23:34:29.437522665 +0000 UTC m=+40.409673999" lastFinishedPulling="2026-04-24 23:34:32.011443926 +0000 UTC m=+42.983595260" observedRunningTime="2026-04-24 23:34:32.349872032 +0000 UTC m=+43.322023376" watchObservedRunningTime="2026-04-24 23:34:32.437843529 +0000 UTC m=+43.409994863" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.311 [INFO][5041] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0 calico-kube-controllers-6cd754fd7f- calico-system 5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b 1002 0 2026-04-24 23:34:04 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6cd754fd7f projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081-3-6-n-e88c017743 calico-kube-controllers-6cd754fd7f-hzjqr eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali92655be1325 [] [] }} ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Namespace="calico-system" Pod="calico-kube-controllers-6cd754fd7f-hzjqr" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.312 [INFO][5041] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Namespace="calico-system" Pod="calico-kube-controllers-6cd754fd7f-hzjqr" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.374 [INFO][5063] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" HandleID="k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.384 [INFO][5063] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" HandleID="k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ef5c0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-e88c017743", "pod":"calico-kube-controllers-6cd754fd7f-hzjqr", "timestamp":"2026-04-24 23:34:32.374642569 +0000 UTC"}, Hostname:"ci-4081-3-6-n-e88c017743", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0002e0f20)} Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.384 [INFO][5063] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.384 [INFO][5063] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.384 [INFO][5063] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-e88c017743' Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.388 [INFO][5063] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.397 [INFO][5063] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.405 [INFO][5063] ipam/ipam.go 526: Trying affinity for 192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.409 [INFO][5063] ipam/ipam.go 160: Attempting to load block cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.411 [INFO][5063] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.411 [INFO][5063] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.115.0/26 handle="k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.414 [INFO][5063] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969 Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.420 [INFO][5063] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.115.0/26 handle="k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.425 [INFO][5063] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.115.5/26] block=192.168.115.0/26 handle="k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.425 [INFO][5063] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.115.5/26] handle="k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.425 [INFO][5063] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:32.460439 containerd[1631]: 2026-04-24 23:34:32.425 [INFO][5063] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.115.5/26] IPv6=[] ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" HandleID="k8s-pod-network.16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.462396 containerd[1631]: 2026-04-24 23:34:32.428 [INFO][5041] cni-plugin/k8s.go 418: Populated endpoint ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Namespace="calico-system" Pod="calico-kube-controllers-6cd754fd7f-hzjqr" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0", GenerateName:"calico-kube-controllers-6cd754fd7f-", Namespace:"calico-system", SelfLink:"", UID:"5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b", ResourceVersion:"1002", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6cd754fd7f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"", Pod:"calico-kube-controllers-6cd754fd7f-hzjqr", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.115.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali92655be1325", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:32.462396 containerd[1631]: 2026-04-24 23:34:32.428 [INFO][5041] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.115.5/32] ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Namespace="calico-system" Pod="calico-kube-controllers-6cd754fd7f-hzjqr" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.462396 containerd[1631]: 2026-04-24 23:34:32.428 [INFO][5041] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali92655be1325 ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Namespace="calico-system" Pod="calico-kube-controllers-6cd754fd7f-hzjqr" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.462396 containerd[1631]: 2026-04-24 23:34:32.430 [INFO][5041] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Namespace="calico-system" Pod="calico-kube-controllers-6cd754fd7f-hzjqr" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.462396 containerd[1631]: 2026-04-24 23:34:32.430 [INFO][5041] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Namespace="calico-system" Pod="calico-kube-controllers-6cd754fd7f-hzjqr" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0", GenerateName:"calico-kube-controllers-6cd754fd7f-", Namespace:"calico-system", SelfLink:"", UID:"5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b", ResourceVersion:"1002", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6cd754fd7f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969", Pod:"calico-kube-controllers-6cd754fd7f-hzjqr", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.115.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali92655be1325", MAC:"16:44:64:54:e8:7e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:32.462396 containerd[1631]: 2026-04-24 23:34:32.440 [INFO][5041] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969" Namespace="calico-system" Pod="calico-kube-controllers-6cd754fd7f-hzjqr" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:32.503941 containerd[1631]: time="2026-04-24T23:34:32.502555266Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:32.503941 containerd[1631]: time="2026-04-24T23:34:32.503368824Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:32.503941 containerd[1631]: time="2026-04-24T23:34:32.503376374Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:32.503941 containerd[1631]: time="2026-04-24T23:34:32.503439274Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:32.602639 systemd-networkd[1262]: cali18f17b6c873: Link UP Apr 24 23:34:32.606276 systemd-networkd[1262]: cali18f17b6c873: Gained carrier Apr 24 23:34:32.638228 containerd[1631]: time="2026-04-24T23:34:32.636746767Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cd754fd7f-hzjqr,Uid:5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b,Namespace:calico-system,Attempt:1,} returns sandbox id \"16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969\"" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.357 [INFO][5052] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0 coredns-674b8bbfcf- kube-system 51e78ef8-95fe-4fb0-adbe-918fd6d5713b 1001 0 2026-04-24 23:33:54 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081-3-6-n-e88c017743 coredns-674b8bbfcf-kc2pn eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali18f17b6c873 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Namespace="kube-system" Pod="coredns-674b8bbfcf-kc2pn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.358 [INFO][5052] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Namespace="kube-system" Pod="coredns-674b8bbfcf-kc2pn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.411 [INFO][5069] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" HandleID="k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.421 [INFO][5069] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" HandleID="k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000277af0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081-3-6-n-e88c017743", "pod":"coredns-674b8bbfcf-kc2pn", "timestamp":"2026-04-24 23:34:32.411492136 +0000 UTC"}, Hostname:"ci-4081-3-6-n-e88c017743", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0002e7080)} Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.421 [INFO][5069] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.425 [INFO][5069] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.425 [INFO][5069] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-e88c017743' Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.493 [INFO][5069] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.505 [INFO][5069] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.518 [INFO][5069] ipam/ipam.go 526: Trying affinity for 192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.523 [INFO][5069] ipam/ipam.go 160: Attempting to load block cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.528 [INFO][5069] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.528 [INFO][5069] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.115.0/26 handle="k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.532 [INFO][5069] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47 Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.538 [INFO][5069] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.115.0/26 handle="k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.554 [INFO][5069] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.115.6/26] block=192.168.115.0/26 handle="k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.554 [INFO][5069] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.115.6/26] handle="k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.554 [INFO][5069] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:32.641641 containerd[1631]: 2026-04-24 23:34:32.554 [INFO][5069] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.115.6/26] IPv6=[] ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" HandleID="k8s-pod-network.91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.642044 containerd[1631]: 2026-04-24 23:34:32.584 [INFO][5052] cni-plugin/k8s.go 418: Populated endpoint ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Namespace="kube-system" Pod="coredns-674b8bbfcf-kc2pn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"51e78ef8-95fe-4fb0-adbe-918fd6d5713b", ResourceVersion:"1001", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 33, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"", Pod:"coredns-674b8bbfcf-kc2pn", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.115.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali18f17b6c873", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:32.642044 containerd[1631]: 2026-04-24 23:34:32.584 [INFO][5052] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.115.6/32] ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Namespace="kube-system" Pod="coredns-674b8bbfcf-kc2pn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.642044 containerd[1631]: 2026-04-24 23:34:32.584 [INFO][5052] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali18f17b6c873 ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Namespace="kube-system" Pod="coredns-674b8bbfcf-kc2pn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.642044 containerd[1631]: 2026-04-24 23:34:32.610 [INFO][5052] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Namespace="kube-system" Pod="coredns-674b8bbfcf-kc2pn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.642044 containerd[1631]: 2026-04-24 23:34:32.621 [INFO][5052] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Namespace="kube-system" Pod="coredns-674b8bbfcf-kc2pn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"51e78ef8-95fe-4fb0-adbe-918fd6d5713b", ResourceVersion:"1001", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 33, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47", Pod:"coredns-674b8bbfcf-kc2pn", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.115.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali18f17b6c873", MAC:"be:80:bf:7e:56:16", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:32.642044 containerd[1631]: 2026-04-24 23:34:32.632 [INFO][5052] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47" Namespace="kube-system" Pod="coredns-674b8bbfcf-kc2pn" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:32.692604 containerd[1631]: time="2026-04-24T23:34:32.692501326Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:32.692604 containerd[1631]: time="2026-04-24T23:34:32.692558736Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:32.692604 containerd[1631]: time="2026-04-24T23:34:32.692567076Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:32.693131 containerd[1631]: time="2026-04-24T23:34:32.692675786Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:32.763903 containerd[1631]: time="2026-04-24T23:34:32.763626206Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-kc2pn,Uid:51e78ef8-95fe-4fb0-adbe-918fd6d5713b,Namespace:kube-system,Attempt:1,} returns sandbox id \"91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47\"" Apr 24 23:34:32.769384 containerd[1631]: time="2026-04-24T23:34:32.769139173Z" level=info msg="CreateContainer within sandbox \"91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 24 23:34:32.778607 containerd[1631]: time="2026-04-24T23:34:32.778537969Z" level=info msg="CreateContainer within sandbox \"91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d00ba7b52594a110c2337c1b6a682910dd260ce2471e31dc579a322e1883dd48\"" Apr 24 23:34:32.779762 containerd[1631]: time="2026-04-24T23:34:32.779002768Z" level=info msg="StartContainer for \"d00ba7b52594a110c2337c1b6a682910dd260ce2471e31dc579a322e1883dd48\"" Apr 24 23:34:32.848331 systemd-networkd[1262]: calief231f26ce3: Gained IPv6LL Apr 24 23:34:32.858691 containerd[1631]: time="2026-04-24T23:34:32.858620277Z" level=info msg="StartContainer for \"d00ba7b52594a110c2337c1b6a682910dd260ce2471e31dc579a322e1883dd48\" returns successfully" Apr 24 23:34:33.118360 containerd[1631]: time="2026-04-24T23:34:33.118306276Z" level=info msg="StopPodSandbox for \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\"" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.159 [INFO][5260] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.160 [INFO][5260] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" iface="eth0" netns="/var/run/netns/cni-0df87044-9bf7-8664-8667-074cec1fac79" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.161 [INFO][5260] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" iface="eth0" netns="/var/run/netns/cni-0df87044-9bf7-8664-8667-074cec1fac79" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.162 [INFO][5260] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" iface="eth0" netns="/var/run/netns/cni-0df87044-9bf7-8664-8667-074cec1fac79" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.162 [INFO][5260] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.162 [INFO][5260] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.186 [INFO][5267] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.187 [INFO][5267] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.187 [INFO][5267] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.191 [WARNING][5267] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.191 [INFO][5267] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.192 [INFO][5267] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:33.196323 containerd[1631]: 2026-04-24 23:34:33.194 [INFO][5260] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:33.197366 containerd[1631]: time="2026-04-24T23:34:33.196784064Z" level=info msg="TearDown network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\" successfully" Apr 24 23:34:33.197366 containerd[1631]: time="2026-04-24T23:34:33.196835054Z" level=info msg="StopPodSandbox for \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\" returns successfully" Apr 24 23:34:33.197844 containerd[1631]: time="2026-04-24T23:34:33.197576992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-l624f,Uid:f187502e-fb21-4086-9792-3a608142290e,Namespace:calico-system,Attempt:1,}" Apr 24 23:34:33.231975 systemd[1]: run-netns-cni\x2d0df87044\x2d9bf7\x2d8664\x2d8667\x2d074cec1fac79.mount: Deactivated successfully. Apr 24 23:34:33.326734 systemd-networkd[1262]: cali8f7acbe7c91: Link UP Apr 24 23:34:33.328063 systemd-networkd[1262]: cali8f7acbe7c91: Gained carrier Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.252 [INFO][5274] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0 goldmane-5b85766d88- calico-system f187502e-fb21-4086-9792-3a608142290e 1020 0 2026-04-24 23:34:03 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:5b85766d88 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4081-3-6-n-e88c017743 goldmane-5b85766d88-l624f eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali8f7acbe7c91 [] [] }} ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Namespace="calico-system" Pod="goldmane-5b85766d88-l624f" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.253 [INFO][5274] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Namespace="calico-system" Pod="goldmane-5b85766d88-l624f" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.292 [INFO][5285] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" HandleID="k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.300 [INFO][5285] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" HandleID="k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002f9570), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-e88c017743", "pod":"goldmane-5b85766d88-l624f", "timestamp":"2026-04-24 23:34:33.292779402 +0000 UTC"}, Hostname:"ci-4081-3-6-n-e88c017743", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00026cdc0)} Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.300 [INFO][5285] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.300 [INFO][5285] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.300 [INFO][5285] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-e88c017743' Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.302 [INFO][5285] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.306 [INFO][5285] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.309 [INFO][5285] ipam/ipam.go 526: Trying affinity for 192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.311 [INFO][5285] ipam/ipam.go 160: Attempting to load block cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.313 [INFO][5285] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.313 [INFO][5285] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.115.0/26 handle="k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.314 [INFO][5285] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2 Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.317 [INFO][5285] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.115.0/26 handle="k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.322 [INFO][5285] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.115.7/26] block=192.168.115.0/26 handle="k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.322 [INFO][5285] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.115.7/26] handle="k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.322 [INFO][5285] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:33.343602 containerd[1631]: 2026-04-24 23:34:33.322 [INFO][5285] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.115.7/26] IPv6=[] ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" HandleID="k8s-pod-network.fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.344319 containerd[1631]: 2026-04-24 23:34:33.324 [INFO][5274] cni-plugin/k8s.go 418: Populated endpoint ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Namespace="calico-system" Pod="goldmane-5b85766d88-l624f" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"f187502e-fb21-4086-9792-3a608142290e", ResourceVersion:"1020", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"", Pod:"goldmane-5b85766d88-l624f", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.115.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali8f7acbe7c91", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:33.344319 containerd[1631]: 2026-04-24 23:34:33.324 [INFO][5274] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.115.7/32] ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Namespace="calico-system" Pod="goldmane-5b85766d88-l624f" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.344319 containerd[1631]: 2026-04-24 23:34:33.324 [INFO][5274] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8f7acbe7c91 ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Namespace="calico-system" Pod="goldmane-5b85766d88-l624f" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.344319 containerd[1631]: 2026-04-24 23:34:33.327 [INFO][5274] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Namespace="calico-system" Pod="goldmane-5b85766d88-l624f" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.344319 containerd[1631]: 2026-04-24 23:34:33.328 [INFO][5274] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Namespace="calico-system" Pod="goldmane-5b85766d88-l624f" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"f187502e-fb21-4086-9792-3a608142290e", ResourceVersion:"1020", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2", Pod:"goldmane-5b85766d88-l624f", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.115.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali8f7acbe7c91", MAC:"f6:d5:52:ec:ea:39", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:33.344319 containerd[1631]: 2026-04-24 23:34:33.340 [INFO][5274] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2" Namespace="calico-system" Pod="goldmane-5b85766d88-l624f" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:33.358911 kubelet[2738]: I0424 23:34:33.358816 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-kc2pn" podStartSLOduration=39.358738529 podStartE2EDuration="39.358738529s" podCreationTimestamp="2026-04-24 23:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-24 23:34:33.355627457 +0000 UTC m=+44.327778791" watchObservedRunningTime="2026-04-24 23:34:33.358738529 +0000 UTC m=+44.330889873" Apr 24 23:34:33.387355 containerd[1631]: time="2026-04-24T23:34:33.387045744Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:33.387355 containerd[1631]: time="2026-04-24T23:34:33.387099624Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:33.387355 containerd[1631]: time="2026-04-24T23:34:33.387107214Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:33.387355 containerd[1631]: time="2026-04-24T23:34:33.387178054Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:33.473804 containerd[1631]: time="2026-04-24T23:34:33.473729253Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-l624f,Uid:f187502e-fb21-4086-9792-3a608142290e,Namespace:calico-system,Attempt:1,} returns sandbox id \"fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2\"" Apr 24 23:34:33.665334 containerd[1631]: time="2026-04-24T23:34:33.665223709Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:33.666130 containerd[1631]: time="2026-04-24T23:34:33.665987998Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.4: active requests=0, bytes read=8792502" Apr 24 23:34:33.668004 containerd[1631]: time="2026-04-24T23:34:33.666929126Z" level=info msg="ImageCreate event name:\"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:33.669112 containerd[1631]: time="2026-04-24T23:34:33.668796261Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:33.669667 containerd[1631]: time="2026-04-24T23:34:33.669338270Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.4\" with image id \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\", size \"10348547\" in 1.657728684s" Apr 24 23:34:33.669667 containerd[1631]: time="2026-04-24T23:34:33.669369990Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\" returns image reference \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\"" Apr 24 23:34:33.670604 containerd[1631]: time="2026-04-24T23:34:33.670591247Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\"" Apr 24 23:34:33.673177 containerd[1631]: time="2026-04-24T23:34:33.673148801Z" level=info msg="CreateContainer within sandbox \"a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Apr 24 23:34:33.687277 containerd[1631]: time="2026-04-24T23:34:33.687242309Z" level=info msg="CreateContainer within sandbox \"a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"4138e0455e8b8f34b814aeda5c53bb38139a98faf1ba1b5ea0e8642b8b3710ac\"" Apr 24 23:34:33.688453 containerd[1631]: time="2026-04-24T23:34:33.687627848Z" level=info msg="StartContainer for \"4138e0455e8b8f34b814aeda5c53bb38139a98faf1ba1b5ea0e8642b8b3710ac\"" Apr 24 23:34:33.744881 containerd[1631]: time="2026-04-24T23:34:33.744838666Z" level=info msg="StartContainer for \"4138e0455e8b8f34b814aeda5c53bb38139a98faf1ba1b5ea0e8642b8b3710ac\" returns successfully" Apr 24 23:34:34.118432 containerd[1631]: time="2026-04-24T23:34:34.117328065Z" level=info msg="StopPodSandbox for \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\"" Apr 24 23:34:34.192891 systemd-networkd[1262]: cali18f17b6c873: Gained IPv6LL Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.190 [INFO][5405] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.190 [INFO][5405] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" iface="eth0" netns="/var/run/netns/cni-4d2bbdda-b8aa-4504-eb7b-3ab4fb922408" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.190 [INFO][5405] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" iface="eth0" netns="/var/run/netns/cni-4d2bbdda-b8aa-4504-eb7b-3ab4fb922408" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.191 [INFO][5405] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" iface="eth0" netns="/var/run/netns/cni-4d2bbdda-b8aa-4504-eb7b-3ab4fb922408" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.191 [INFO][5405] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.191 [INFO][5405] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.222 [INFO][5412] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.223 [INFO][5412] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.223 [INFO][5412] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.227 [WARNING][5412] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.227 [INFO][5412] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.228 [INFO][5412] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:34.232625 containerd[1631]: 2026-04-24 23:34:34.230 [INFO][5405] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:34.235454 containerd[1631]: time="2026-04-24T23:34:34.235213635Z" level=info msg="TearDown network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\" successfully" Apr 24 23:34:34.235454 containerd[1631]: time="2026-04-24T23:34:34.235243075Z" level=info msg="StopPodSandbox for \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\" returns successfully" Apr 24 23:34:34.236165 containerd[1631]: time="2026-04-24T23:34:34.235913004Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85f77d67c7-7k5mt,Uid:c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b,Namespace:calico-system,Attempt:1,}" Apr 24 23:34:34.237466 systemd[1]: run-netns-cni\x2d4d2bbdda\x2db8aa\x2d4504\x2deb7b\x2d3ab4fb922408.mount: Deactivated successfully. Apr 24 23:34:34.329640 systemd-networkd[1262]: cali5216d3d4b21: Link UP Apr 24 23:34:34.330213 systemd-networkd[1262]: cali5216d3d4b21: Gained carrier Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.274 [INFO][5426] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0 calico-apiserver-85f77d67c7- calico-system c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b 1045 0 2026-04-24 23:34:03 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:85f77d67c7 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081-3-6-n-e88c017743 calico-apiserver-85f77d67c7-7k5mt eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali5216d3d4b21 [] [] }} ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-7k5mt" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.275 [INFO][5426] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-7k5mt" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.296 [INFO][5435] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" HandleID="k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.301 [INFO][5435] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" HandleID="k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002f74c0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-e88c017743", "pod":"calico-apiserver-85f77d67c7-7k5mt", "timestamp":"2026-04-24 23:34:34.296631815 +0000 UTC"}, Hostname:"ci-4081-3-6-n-e88c017743", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00026edc0)} Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.301 [INFO][5435] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.301 [INFO][5435] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.301 [INFO][5435] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-e88c017743' Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.303 [INFO][5435] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.306 [INFO][5435] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.309 [INFO][5435] ipam/ipam.go 526: Trying affinity for 192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.311 [INFO][5435] ipam/ipam.go 160: Attempting to load block cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.312 [INFO][5435] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.115.0/26 host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.312 [INFO][5435] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.115.0/26 handle="k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.314 [INFO][5435] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.317 [INFO][5435] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.115.0/26 handle="k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.323 [INFO][5435] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.115.8/26] block=192.168.115.0/26 handle="k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.323 [INFO][5435] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.115.8/26] handle="k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" host="ci-4081-3-6-n-e88c017743" Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.323 [INFO][5435] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:34.343721 containerd[1631]: 2026-04-24 23:34:34.323 [INFO][5435] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.115.8/26] IPv6=[] ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" HandleID="k8s-pod-network.2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.344409 containerd[1631]: 2026-04-24 23:34:34.325 [INFO][5426] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-7k5mt" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0", GenerateName:"calico-apiserver-85f77d67c7-", Namespace:"calico-system", SelfLink:"", UID:"c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b", ResourceVersion:"1045", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85f77d67c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"", Pod:"calico-apiserver-85f77d67c7-7k5mt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.115.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali5216d3d4b21", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:34.344409 containerd[1631]: 2026-04-24 23:34:34.325 [INFO][5426] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.115.8/32] ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-7k5mt" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.344409 containerd[1631]: 2026-04-24 23:34:34.325 [INFO][5426] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5216d3d4b21 ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-7k5mt" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.344409 containerd[1631]: 2026-04-24 23:34:34.329 [INFO][5426] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-7k5mt" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.344409 containerd[1631]: 2026-04-24 23:34:34.329 [INFO][5426] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-7k5mt" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0", GenerateName:"calico-apiserver-85f77d67c7-", Namespace:"calico-system", SelfLink:"", UID:"c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b", ResourceVersion:"1045", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85f77d67c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af", Pod:"calico-apiserver-85f77d67c7-7k5mt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.115.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali5216d3d4b21", MAC:"a6:10:34:e3:f1:a9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:34.344409 containerd[1631]: 2026-04-24 23:34:34.339 [INFO][5426] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af" Namespace="calico-system" Pod="calico-apiserver-85f77d67c7-7k5mt" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:34.380037 containerd[1631]: time="2026-04-24T23:34:34.379719560Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 24 23:34:34.380037 containerd[1631]: time="2026-04-24T23:34:34.379758170Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 24 23:34:34.380037 containerd[1631]: time="2026-04-24T23:34:34.379768390Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:34.380848 containerd[1631]: time="2026-04-24T23:34:34.380059069Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 24 23:34:34.387164 systemd-networkd[1262]: cali92655be1325: Gained IPv6LL Apr 24 23:34:34.496848 containerd[1631]: time="2026-04-24T23:34:34.496717811Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85f77d67c7-7k5mt,Uid:c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b,Namespace:calico-system,Attempt:1,} returns sandbox id \"2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af\"" Apr 24 23:34:34.503479 containerd[1631]: time="2026-04-24T23:34:34.503188128Z" level=info msg="CreateContainer within sandbox \"2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 24 23:34:34.514628 containerd[1631]: time="2026-04-24T23:34:34.514607414Z" level=info msg="CreateContainer within sandbox \"2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"5d1b5d668d9cd79422c80aac4b47b68fb1260bf3e9f5dc19c8a48a232ccb6ed0\"" Apr 24 23:34:34.516211 containerd[1631]: time="2026-04-24T23:34:34.515224783Z" level=info msg="StartContainer for \"5d1b5d668d9cd79422c80aac4b47b68fb1260bf3e9f5dc19c8a48a232ccb6ed0\"" Apr 24 23:34:34.577200 systemd-networkd[1262]: cali8f7acbe7c91: Gained IPv6LL Apr 24 23:34:34.611969 containerd[1631]: time="2026-04-24T23:34:34.611482989Z" level=info msg="StartContainer for \"5d1b5d668d9cd79422c80aac4b47b68fb1260bf3e9f5dc19c8a48a232ccb6ed0\" returns successfully" Apr 24 23:34:35.392779 kubelet[2738]: I0424 23:34:35.390687 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-85f77d67c7-7k5mt" podStartSLOduration=32.39063256 podStartE2EDuration="32.39063256s" podCreationTimestamp="2026-04-24 23:34:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-24 23:34:35.390286031 +0000 UTC m=+46.362437375" watchObservedRunningTime="2026-04-24 23:34:35.39063256 +0000 UTC m=+46.362783894" Apr 24 23:34:35.536929 systemd-networkd[1262]: cali5216d3d4b21: Gained IPv6LL Apr 24 23:34:36.285200 containerd[1631]: time="2026-04-24T23:34:36.284632180Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:36.285745 containerd[1631]: time="2026-04-24T23:34:36.285717799Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.4: active requests=0, bytes read=52406348" Apr 24 23:34:36.287788 containerd[1631]: time="2026-04-24T23:34:36.287757674Z" level=info msg="ImageCreate event name:\"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:36.290616 containerd[1631]: time="2026-04-24T23:34:36.290598490Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:36.291621 containerd[1631]: time="2026-04-24T23:34:36.291529577Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" with image id \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\", size \"53962361\" in 2.62084871s" Apr 24 23:34:36.291621 containerd[1631]: time="2026-04-24T23:34:36.291589567Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" returns image reference \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\"" Apr 24 23:34:36.293157 containerd[1631]: time="2026-04-24T23:34:36.293005596Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\"" Apr 24 23:34:36.318388 containerd[1631]: time="2026-04-24T23:34:36.318347671Z" level=info msg="CreateContainer within sandbox \"16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Apr 24 23:34:36.344860 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2883740102.mount: Deactivated successfully. Apr 24 23:34:36.345438 containerd[1631]: time="2026-04-24T23:34:36.345399463Z" level=info msg="CreateContainer within sandbox \"16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"f49693d76d271b52c256b698c9428ab82f55700cb8db9e6dfb5c2f74e9220be0\"" Apr 24 23:34:36.347153 containerd[1631]: time="2026-04-24T23:34:36.346832960Z" level=info msg="StartContainer for \"f49693d76d271b52c256b698c9428ab82f55700cb8db9e6dfb5c2f74e9220be0\"" Apr 24 23:34:36.455742 containerd[1631]: time="2026-04-24T23:34:36.455545359Z" level=info msg="StartContainer for \"f49693d76d271b52c256b698c9428ab82f55700cb8db9e6dfb5c2f74e9220be0\" returns successfully" Apr 24 23:34:37.409224 kubelet[2738]: I0424 23:34:37.406702 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-6cd754fd7f-hzjqr" podStartSLOduration=29.755429628 podStartE2EDuration="33.406678079s" podCreationTimestamp="2026-04-24 23:34:04 +0000 UTC" firstStartedPulling="2026-04-24 23:34:32.641167695 +0000 UTC m=+43.613319029" lastFinishedPulling="2026-04-24 23:34:36.292416136 +0000 UTC m=+47.264567480" observedRunningTime="2026-04-24 23:34:37.405311731 +0000 UTC m=+48.377463105" watchObservedRunningTime="2026-04-24 23:34:37.406678079 +0000 UTC m=+48.378829453" Apr 24 23:34:38.372585 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3714949094.mount: Deactivated successfully. Apr 24 23:34:38.670185 containerd[1631]: time="2026-04-24T23:34:38.670043552Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:38.671269 containerd[1631]: time="2026-04-24T23:34:38.671180011Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.4: active requests=0, bytes read=55623386" Apr 24 23:34:38.673503 containerd[1631]: time="2026-04-24T23:34:38.672331599Z" level=info msg="ImageCreate event name:\"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:38.675123 containerd[1631]: time="2026-04-24T23:34:38.674501196Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:38.675123 containerd[1631]: time="2026-04-24T23:34:38.675001435Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" with image id \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\", size \"55623232\" in 2.381977419s" Apr 24 23:34:38.675123 containerd[1631]: time="2026-04-24T23:34:38.675025455Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" returns image reference \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\"" Apr 24 23:34:38.676108 containerd[1631]: time="2026-04-24T23:34:38.676094233Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\"" Apr 24 23:34:38.678596 containerd[1631]: time="2026-04-24T23:34:38.678567330Z" level=info msg="CreateContainer within sandbox \"fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Apr 24 23:34:38.688886 containerd[1631]: time="2026-04-24T23:34:38.688847194Z" level=info msg="CreateContainer within sandbox \"fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"203ba8e79a4d43760a778e6b37541469774f0c2f9a256d4b8cf3c6f2cc3f3156\"" Apr 24 23:34:38.689656 containerd[1631]: time="2026-04-24T23:34:38.689247204Z" level=info msg="StartContainer for \"203ba8e79a4d43760a778e6b37541469774f0c2f9a256d4b8cf3c6f2cc3f3156\"" Apr 24 23:34:38.757755 containerd[1631]: time="2026-04-24T23:34:38.757679805Z" level=info msg="StartContainer for \"203ba8e79a4d43760a778e6b37541469774f0c2f9a256d4b8cf3c6f2cc3f3156\" returns successfully" Apr 24 23:34:39.412591 kubelet[2738]: I0424 23:34:39.412489 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-5b85766d88-l624f" podStartSLOduration=31.212677109 podStartE2EDuration="36.412475696s" podCreationTimestamp="2026-04-24 23:34:03 +0000 UTC" firstStartedPulling="2026-04-24 23:34:33.476192487 +0000 UTC m=+44.448343821" lastFinishedPulling="2026-04-24 23:34:38.675991064 +0000 UTC m=+49.648142408" observedRunningTime="2026-04-24 23:34:39.411644908 +0000 UTC m=+50.383796242" watchObservedRunningTime="2026-04-24 23:34:39.412475696 +0000 UTC m=+50.384627040" Apr 24 23:34:40.683029 containerd[1631]: time="2026-04-24T23:34:40.682987821Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:40.686099 containerd[1631]: time="2026-04-24T23:34:40.685194748Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4: active requests=0, bytes read=14704317" Apr 24 23:34:40.686099 containerd[1631]: time="2026-04-24T23:34:40.685887847Z" level=info msg="ImageCreate event name:\"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:40.690405 containerd[1631]: time="2026-04-24T23:34:40.689919473Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 24 23:34:40.691269 containerd[1631]: time="2026-04-24T23:34:40.691247221Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" with image id \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\", size \"16260314\" in 2.014336869s" Apr 24 23:34:40.691310 containerd[1631]: time="2026-04-24T23:34:40.691272571Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" returns image reference \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\"" Apr 24 23:34:40.697683 containerd[1631]: time="2026-04-24T23:34:40.697576103Z" level=info msg="CreateContainer within sandbox \"a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Apr 24 23:34:40.718115 containerd[1631]: time="2026-04-24T23:34:40.717236590Z" level=info msg="CreateContainer within sandbox \"a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"ebc5b09eb6d14d5b6ea2f96369c1942db3254f2dbb2b6d6e02a70a4b975464f6\"" Apr 24 23:34:40.719160 containerd[1631]: time="2026-04-24T23:34:40.719140948Z" level=info msg="StartContainer for \"ebc5b09eb6d14d5b6ea2f96369c1942db3254f2dbb2b6d6e02a70a4b975464f6\"" Apr 24 23:34:40.821275 containerd[1631]: time="2026-04-24T23:34:40.819855820Z" level=info msg="StartContainer for \"ebc5b09eb6d14d5b6ea2f96369c1942db3254f2dbb2b6d6e02a70a4b975464f6\" returns successfully" Apr 24 23:34:41.203493 kubelet[2738]: I0424 23:34:41.203349 2738 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Apr 24 23:34:41.205737 kubelet[2738]: I0424 23:34:41.205703 2738 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Apr 24 23:34:49.118385 containerd[1631]: time="2026-04-24T23:34:49.118298844Z" level=info msg="StopPodSandbox for \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\"" Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.188 [WARNING][5818] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"51e78ef8-95fe-4fb0-adbe-918fd6d5713b", ResourceVersion:"1027", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 33, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47", Pod:"coredns-674b8bbfcf-kc2pn", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.115.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali18f17b6c873", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.188 [INFO][5818] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.188 [INFO][5818] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" iface="eth0" netns="" Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.188 [INFO][5818] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.188 [INFO][5818] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.214 [INFO][5826] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.214 [INFO][5826] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.214 [INFO][5826] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.221 [WARNING][5826] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.221 [INFO][5826] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.222 [INFO][5826] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.228535 containerd[1631]: 2026-04-24 23:34:49.225 [INFO][5818] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:49.228535 containerd[1631]: time="2026-04-24T23:34:49.228450703Z" level=info msg="TearDown network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\" successfully" Apr 24 23:34:49.228535 containerd[1631]: time="2026-04-24T23:34:49.228471293Z" level=info msg="StopPodSandbox for \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\" returns successfully" Apr 24 23:34:49.229218 containerd[1631]: time="2026-04-24T23:34:49.229194892Z" level=info msg="RemovePodSandbox for \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\"" Apr 24 23:34:49.229218 containerd[1631]: time="2026-04-24T23:34:49.229218622Z" level=info msg="Forcibly stopping sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\"" Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.252 [WARNING][5841] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"51e78ef8-95fe-4fb0-adbe-918fd6d5713b", ResourceVersion:"1027", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 33, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"91125fb98748f578799373a567918bcdc9a6ba8996801d3bf983b4b635ab3c47", Pod:"coredns-674b8bbfcf-kc2pn", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.115.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali18f17b6c873", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.253 [INFO][5841] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.253 [INFO][5841] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" iface="eth0" netns="" Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.253 [INFO][5841] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.253 [INFO][5841] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.267 [INFO][5848] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.267 [INFO][5848] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.267 [INFO][5848] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.271 [WARNING][5848] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.271 [INFO][5848] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" HandleID="k8s-pod-network.76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--kc2pn-eth0" Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.272 [INFO][5848] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.276072 containerd[1631]: 2026-04-24 23:34:49.274 [INFO][5841] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29" Apr 24 23:34:49.276414 containerd[1631]: time="2026-04-24T23:34:49.276120898Z" level=info msg="TearDown network for sandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\" successfully" Apr 24 23:34:49.280268 containerd[1631]: time="2026-04-24T23:34:49.280135608Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 24 23:34:49.280268 containerd[1631]: time="2026-04-24T23:34:49.280183778Z" level=info msg="RemovePodSandbox \"76a30af52aa083a327760204ece9b43d907c50c4293fc341dbe3260f2cbb4c29\" returns successfully" Apr 24 23:34:49.281340 containerd[1631]: time="2026-04-24T23:34:49.281109798Z" level=info msg="StopPodSandbox for \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\"" Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.304 [WARNING][5863] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0", GenerateName:"calico-kube-controllers-6cd754fd7f-", Namespace:"calico-system", SelfLink:"", UID:"5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b", ResourceVersion:"1086", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6cd754fd7f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969", Pod:"calico-kube-controllers-6cd754fd7f-hzjqr", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.115.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali92655be1325", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.305 [INFO][5863] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.305 [INFO][5863] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" iface="eth0" netns="" Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.305 [INFO][5863] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.305 [INFO][5863] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.319 [INFO][5871] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.319 [INFO][5871] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.319 [INFO][5871] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.324 [WARNING][5871] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.324 [INFO][5871] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.325 [INFO][5871] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.329911 containerd[1631]: 2026-04-24 23:34:49.327 [INFO][5863] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:49.330349 containerd[1631]: time="2026-04-24T23:34:49.329950853Z" level=info msg="TearDown network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\" successfully" Apr 24 23:34:49.330349 containerd[1631]: time="2026-04-24T23:34:49.329969923Z" level=info msg="StopPodSandbox for \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\" returns successfully" Apr 24 23:34:49.330526 containerd[1631]: time="2026-04-24T23:34:49.330413704Z" level=info msg="RemovePodSandbox for \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\"" Apr 24 23:34:49.330553 containerd[1631]: time="2026-04-24T23:34:49.330529433Z" level=info msg="Forcibly stopping sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\"" Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.359 [WARNING][5885] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0", GenerateName:"calico-kube-controllers-6cd754fd7f-", Namespace:"calico-system", SelfLink:"", UID:"5a5df8c3-bcf4-4b8d-8bc6-874c87b1775b", ResourceVersion:"1086", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6cd754fd7f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"16b7b0289e886da41b200154782bcddea77352b22b75d0a29919a581bfe2b969", Pod:"calico-kube-controllers-6cd754fd7f-hzjqr", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.115.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali92655be1325", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.359 [INFO][5885] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.359 [INFO][5885] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" iface="eth0" netns="" Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.359 [INFO][5885] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.359 [INFO][5885] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.374 [INFO][5892] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.374 [INFO][5892] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.374 [INFO][5892] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.378 [WARNING][5892] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.378 [INFO][5892] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" HandleID="k8s-pod-network.84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--kube--controllers--6cd754fd7f--hzjqr-eth0" Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.379 [INFO][5892] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.382933 containerd[1631]: 2026-04-24 23:34:49.381 [INFO][5885] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde" Apr 24 23:34:49.382933 containerd[1631]: time="2026-04-24T23:34:49.382897928Z" level=info msg="TearDown network for sandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\" successfully" Apr 24 23:34:49.386590 containerd[1631]: time="2026-04-24T23:34:49.386565877Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 24 23:34:49.386650 containerd[1631]: time="2026-04-24T23:34:49.386607407Z" level=info msg="RemovePodSandbox \"84ef6e18c2fefda2abd690c92f38f3e5b42057bdbb85149671e3ba98b6a46fde\" returns successfully" Apr 24 23:34:49.387071 containerd[1631]: time="2026-04-24T23:34:49.387036167Z" level=info msg="StopPodSandbox for \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\"" Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.410 [WARNING][5906] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0", GenerateName:"calico-apiserver-85f77d67c7-", Namespace:"calico-system", SelfLink:"", UID:"c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b", ResourceVersion:"1065", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85f77d67c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af", Pod:"calico-apiserver-85f77d67c7-7k5mt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.115.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali5216d3d4b21", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.410 [INFO][5906] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.410 [INFO][5906] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" iface="eth0" netns="" Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.410 [INFO][5906] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.410 [INFO][5906] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.424 [INFO][5914] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.424 [INFO][5914] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.425 [INFO][5914] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.430 [WARNING][5914] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.430 [INFO][5914] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.431 [INFO][5914] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.434535 containerd[1631]: 2026-04-24 23:34:49.432 [INFO][5906] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:49.434939 containerd[1631]: time="2026-04-24T23:34:49.434564293Z" level=info msg="TearDown network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\" successfully" Apr 24 23:34:49.434939 containerd[1631]: time="2026-04-24T23:34:49.434585393Z" level=info msg="StopPodSandbox for \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\" returns successfully" Apr 24 23:34:49.434939 containerd[1631]: time="2026-04-24T23:34:49.434918883Z" level=info msg="RemovePodSandbox for \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\"" Apr 24 23:34:49.434939 containerd[1631]: time="2026-04-24T23:34:49.434936703Z" level=info msg="Forcibly stopping sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\"" Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.460 [WARNING][5928] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0", GenerateName:"calico-apiserver-85f77d67c7-", Namespace:"calico-system", SelfLink:"", UID:"c864c683-ffc2-4ce1-ad8a-8aa9e2004a8b", ResourceVersion:"1065", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85f77d67c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"2fbbaecbb0f761abcec0b775e72004e4daaa36a07eedb532deb9e2e51a5c70af", Pod:"calico-apiserver-85f77d67c7-7k5mt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.115.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali5216d3d4b21", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.460 [INFO][5928] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.460 [INFO][5928] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" iface="eth0" netns="" Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.460 [INFO][5928] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.460 [INFO][5928] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.474 [INFO][5936] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.474 [INFO][5936] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.474 [INFO][5936] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.478 [WARNING][5936] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.478 [INFO][5936] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" HandleID="k8s-pod-network.3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--7k5mt-eth0" Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.480 [INFO][5936] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.483430 containerd[1631]: 2026-04-24 23:34:49.481 [INFO][5928] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f" Apr 24 23:34:49.483430 containerd[1631]: time="2026-04-24T23:34:49.483330479Z" level=info msg="TearDown network for sandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\" successfully" Apr 24 23:34:49.489679 containerd[1631]: time="2026-04-24T23:34:49.489651957Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 24 23:34:49.489726 containerd[1631]: time="2026-04-24T23:34:49.489686127Z" level=info msg="RemovePodSandbox \"3881f0bde0159c0041b8d8b02ae266d4979b039480f9fa02b4fcc32b857f040f\" returns successfully" Apr 24 23:34:49.489938 containerd[1631]: time="2026-04-24T23:34:49.489919238Z" level=info msg="StopPodSandbox for \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\"" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.523 [WARNING][5951] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.523 [INFO][5951] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.523 [INFO][5951] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" iface="eth0" netns="" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.523 [INFO][5951] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.523 [INFO][5951] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.538 [INFO][5958] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.538 [INFO][5958] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.538 [INFO][5958] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.543 [WARNING][5958] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.543 [INFO][5958] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.544 [INFO][5958] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.547520 containerd[1631]: 2026-04-24 23:34:49.545 [INFO][5951] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:49.547819 containerd[1631]: time="2026-04-24T23:34:49.547570701Z" level=info msg="TearDown network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\" successfully" Apr 24 23:34:49.547819 containerd[1631]: time="2026-04-24T23:34:49.547605561Z" level=info msg="StopPodSandbox for \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\" returns successfully" Apr 24 23:34:49.548021 containerd[1631]: time="2026-04-24T23:34:49.548001261Z" level=info msg="RemovePodSandbox for \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\"" Apr 24 23:34:49.548042 containerd[1631]: time="2026-04-24T23:34:49.548024521Z" level=info msg="Forcibly stopping sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\"" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.571 [WARNING][5973] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" WorkloadEndpoint="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.571 [INFO][5973] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.571 [INFO][5973] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" iface="eth0" netns="" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.571 [INFO][5973] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.571 [INFO][5973] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.586 [INFO][5980] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.586 [INFO][5980] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.586 [INFO][5980] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.590 [WARNING][5980] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.590 [INFO][5980] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" HandleID="k8s-pod-network.41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Workload="ci--4081--3--6--n--e88c017743-k8s-whisker--54cb6594cd--g4bq8-eth0" Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.591 [INFO][5980] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.594250 containerd[1631]: 2026-04-24 23:34:49.592 [INFO][5973] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a" Apr 24 23:34:49.595363 containerd[1631]: time="2026-04-24T23:34:49.594320837Z" level=info msg="TearDown network for sandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\" successfully" Apr 24 23:34:49.598190 containerd[1631]: time="2026-04-24T23:34:49.598149606Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 24 23:34:49.598241 containerd[1631]: time="2026-04-24T23:34:49.598192956Z" level=info msg="RemovePodSandbox \"41cdf7dc24fe346f808e4f4b740fbac1ae29c8b71f9b6ff4a3cad8767b240f2a\" returns successfully" Apr 24 23:34:49.598610 containerd[1631]: time="2026-04-24T23:34:49.598585106Z" level=info msg="StopPodSandbox for \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\"" Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.624 [WARNING][5994] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0", GenerateName:"calico-apiserver-85f77d67c7-", Namespace:"calico-system", SelfLink:"", UID:"0887e740-8164-4423-b084-a17d88aff0ef", ResourceVersion:"1033", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85f77d67c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82", Pod:"calico-apiserver-85f77d67c7-lz87d", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.115.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali9b95f792bb7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.624 [INFO][5994] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.624 [INFO][5994] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" iface="eth0" netns="" Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.624 [INFO][5994] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.624 [INFO][5994] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.639 [INFO][6003] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.639 [INFO][6003] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.639 [INFO][6003] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.645 [WARNING][6003] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.645 [INFO][6003] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.646 [INFO][6003] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.651459 containerd[1631]: 2026-04-24 23:34:49.649 [INFO][5994] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:49.652641 containerd[1631]: time="2026-04-24T23:34:49.651830600Z" level=info msg="TearDown network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\" successfully" Apr 24 23:34:49.652641 containerd[1631]: time="2026-04-24T23:34:49.651853290Z" level=info msg="StopPodSandbox for \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\" returns successfully" Apr 24 23:34:49.652641 containerd[1631]: time="2026-04-24T23:34:49.652467290Z" level=info msg="RemovePodSandbox for \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\"" Apr 24 23:34:49.652641 containerd[1631]: time="2026-04-24T23:34:49.652504540Z" level=info msg="Forcibly stopping sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\"" Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.674 [WARNING][6018] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0", GenerateName:"calico-apiserver-85f77d67c7-", Namespace:"calico-system", SelfLink:"", UID:"0887e740-8164-4423-b084-a17d88aff0ef", ResourceVersion:"1033", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85f77d67c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"d40c95daf24eb6ecce945f6586b21667f03a276e88bdadf1a429da178f164e82", Pod:"calico-apiserver-85f77d67c7-lz87d", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.115.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali9b95f792bb7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.675 [INFO][6018] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.675 [INFO][6018] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" iface="eth0" netns="" Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.675 [INFO][6018] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.676 [INFO][6018] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.691 [INFO][6025] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.691 [INFO][6025] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.691 [INFO][6025] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.695 [WARNING][6025] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.695 [INFO][6025] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" HandleID="k8s-pod-network.71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Workload="ci--4081--3--6--n--e88c017743-k8s-calico--apiserver--85f77d67c7--lz87d-eth0" Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.696 [INFO][6025] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.700165 containerd[1631]: 2026-04-24 23:34:49.698 [INFO][6018] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20" Apr 24 23:34:49.700165 containerd[1631]: time="2026-04-24T23:34:49.699749986Z" level=info msg="TearDown network for sandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\" successfully" Apr 24 23:34:49.703693 containerd[1631]: time="2026-04-24T23:34:49.703658986Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 24 23:34:49.703739 containerd[1631]: time="2026-04-24T23:34:49.703708436Z" level=info msg="RemovePodSandbox \"71ecd34a0937a702a6671d9cc92040a98ef18ecfcfb197160af1227e5fc93d20\" returns successfully" Apr 24 23:34:49.704236 containerd[1631]: time="2026-04-24T23:34:49.704157646Z" level=info msg="StopPodSandbox for \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\"" Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.729 [WARNING][6040] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"d8da1862-fd17-4f27-a872-feaa69b7162e", ResourceVersion:"987", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 33, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9", Pod:"coredns-674b8bbfcf-2gd6v", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.115.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali02ca9128171", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.729 [INFO][6040] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.729 [INFO][6040] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" iface="eth0" netns="" Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.729 [INFO][6040] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.729 [INFO][6040] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.746 [INFO][6048] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.746 [INFO][6048] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.746 [INFO][6048] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.750 [WARNING][6048] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.750 [INFO][6048] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.751 [INFO][6048] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.755115 containerd[1631]: 2026-04-24 23:34:49.753 [INFO][6040] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:49.755566 containerd[1631]: time="2026-04-24T23:34:49.755211231Z" level=info msg="TearDown network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\" successfully" Apr 24 23:34:49.755566 containerd[1631]: time="2026-04-24T23:34:49.755232151Z" level=info msg="StopPodSandbox for \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\" returns successfully" Apr 24 23:34:49.755922 containerd[1631]: time="2026-04-24T23:34:49.755699601Z" level=info msg="RemovePodSandbox for \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\"" Apr 24 23:34:49.755922 containerd[1631]: time="2026-04-24T23:34:49.755717791Z" level=info msg="Forcibly stopping sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\"" Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.780 [WARNING][6062] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"d8da1862-fd17-4f27-a872-feaa69b7162e", ResourceVersion:"987", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 33, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"633b2e71d6118e0dbd9d90dcc9b6a913b94d2eccb3d7a8231d73b7d1b3a795e9", Pod:"coredns-674b8bbfcf-2gd6v", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.115.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali02ca9128171", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.781 [INFO][6062] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.781 [INFO][6062] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" iface="eth0" netns="" Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.781 [INFO][6062] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.781 [INFO][6062] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.795 [INFO][6069] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.795 [INFO][6069] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.795 [INFO][6069] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.799 [WARNING][6069] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.799 [INFO][6069] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" HandleID="k8s-pod-network.8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Workload="ci--4081--3--6--n--e88c017743-k8s-coredns--674b8bbfcf--2gd6v-eth0" Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.800 [INFO][6069] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.803372 containerd[1631]: 2026-04-24 23:34:49.801 [INFO][6062] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb" Apr 24 23:34:49.803743 containerd[1631]: time="2026-04-24T23:34:49.803459847Z" level=info msg="TearDown network for sandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\" successfully" Apr 24 23:34:49.807804 containerd[1631]: time="2026-04-24T23:34:49.807583286Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 24 23:34:49.807804 containerd[1631]: time="2026-04-24T23:34:49.807672555Z" level=info msg="RemovePodSandbox \"8453c2cd7e4b7f2ec182a87a1fa7b47705451e712ee54461162e819817d68aeb\" returns successfully" Apr 24 23:34:49.808590 containerd[1631]: time="2026-04-24T23:34:49.808345406Z" level=info msg="StopPodSandbox for \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\"" Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.831 [WARNING][6083] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"f187502e-fb21-4086-9792-3a608142290e", ResourceVersion:"1099", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2", Pod:"goldmane-5b85766d88-l624f", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.115.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali8f7acbe7c91", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.831 [INFO][6083] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.832 [INFO][6083] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" iface="eth0" netns="" Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.832 [INFO][6083] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.832 [INFO][6083] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.848 [INFO][6090] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.848 [INFO][6090] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.848 [INFO][6090] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.852 [WARNING][6090] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.852 [INFO][6090] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.853 [INFO][6090] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.857131 containerd[1631]: 2026-04-24 23:34:49.855 [INFO][6083] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:49.857915 containerd[1631]: time="2026-04-24T23:34:49.857160401Z" level=info msg="TearDown network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\" successfully" Apr 24 23:34:49.857915 containerd[1631]: time="2026-04-24T23:34:49.857180081Z" level=info msg="StopPodSandbox for \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\" returns successfully" Apr 24 23:34:49.857915 containerd[1631]: time="2026-04-24T23:34:49.857538692Z" level=info msg="RemovePodSandbox for \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\"" Apr 24 23:34:49.857915 containerd[1631]: time="2026-04-24T23:34:49.857562092Z" level=info msg="Forcibly stopping sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\"" Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.881 [WARNING][6104] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"f187502e-fb21-4086-9792-3a608142290e", ResourceVersion:"1099", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"fe40183e5aed2976423835110c688d75119fc36f9e481b24393a2453b3ea61a2", Pod:"goldmane-5b85766d88-l624f", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.115.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali8f7acbe7c91", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.881 [INFO][6104] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.881 [INFO][6104] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" iface="eth0" netns="" Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.881 [INFO][6104] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.881 [INFO][6104] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.897 [INFO][6111] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.897 [INFO][6111] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.897 [INFO][6111] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.901 [WARNING][6111] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.901 [INFO][6111] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" HandleID="k8s-pod-network.3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Workload="ci--4081--3--6--n--e88c017743-k8s-goldmane--5b85766d88--l624f-eth0" Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.902 [INFO][6111] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.905465 containerd[1631]: 2026-04-24 23:34:49.903 [INFO][6104] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f" Apr 24 23:34:49.905465 containerd[1631]: time="2026-04-24T23:34:49.905440318Z" level=info msg="TearDown network for sandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\" successfully" Apr 24 23:34:49.909735 containerd[1631]: time="2026-04-24T23:34:49.909692596Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 24 23:34:49.909797 containerd[1631]: time="2026-04-24T23:34:49.909771156Z" level=info msg="RemovePodSandbox \"3c239cbb94f34677955f7fa4f6cbf39ee4e02a5cf93934fb9dbd0891b61aba2f\" returns successfully" Apr 24 23:34:49.910434 containerd[1631]: time="2026-04-24T23:34:49.910225506Z" level=info msg="StopPodSandbox for \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\"" Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.935 [WARNING][6126] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7556f2ac-97a1-4b30-8823-9a8d6f3243fe", ResourceVersion:"1126", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b", Pod:"csi-node-driver-kflpl", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.115.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calief231f26ce3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.936 [INFO][6126] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.936 [INFO][6126] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" iface="eth0" netns="" Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.936 [INFO][6126] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.936 [INFO][6126] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.954 [INFO][6134] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.954 [INFO][6134] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.954 [INFO][6134] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.959 [WARNING][6134] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.959 [INFO][6134] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.960 [INFO][6134] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:49.964606 containerd[1631]: 2026-04-24 23:34:49.962 [INFO][6126] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:49.965001 containerd[1631]: time="2026-04-24T23:34:49.964611351Z" level=info msg="TearDown network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\" successfully" Apr 24 23:34:49.965001 containerd[1631]: time="2026-04-24T23:34:49.964628820Z" level=info msg="StopPodSandbox for \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\" returns successfully" Apr 24 23:34:49.965258 containerd[1631]: time="2026-04-24T23:34:49.965215470Z" level=info msg="RemovePodSandbox for \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\"" Apr 24 23:34:49.965258 containerd[1631]: time="2026-04-24T23:34:49.965236150Z" level=info msg="Forcibly stopping sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\"" Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:49.988 [WARNING][6148] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7556f2ac-97a1-4b30-8823-9a8d6f3243fe", ResourceVersion:"1126", Generation:0, CreationTimestamp:time.Date(2026, time.April, 24, 23, 34, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-e88c017743", ContainerID:"a733ae11eb8c5017bc09a18843257d0b8bfce626a395b1626ff4baa9fc9d0b9b", Pod:"csi-node-driver-kflpl", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.115.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calief231f26ce3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:49.988 [INFO][6148] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:49.988 [INFO][6148] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" iface="eth0" netns="" Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:49.988 [INFO][6148] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:49.988 [INFO][6148] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:50.003 [INFO][6155] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:50.003 [INFO][6155] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:50.003 [INFO][6155] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:50.007 [WARNING][6155] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:50.007 [INFO][6155] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" HandleID="k8s-pod-network.04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Workload="ci--4081--3--6--n--e88c017743-k8s-csi--node--driver--kflpl-eth0" Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:50.008 [INFO][6155] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 24 23:34:50.012248 containerd[1631]: 2026-04-24 23:34:50.010 [INFO][6148] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e" Apr 24 23:34:50.012573 containerd[1631]: time="2026-04-24T23:34:50.012275388Z" level=info msg="TearDown network for sandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\" successfully" Apr 24 23:34:50.015693 containerd[1631]: time="2026-04-24T23:34:50.015667397Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 24 23:34:50.015752 containerd[1631]: time="2026-04-24T23:34:50.015721747Z" level=info msg="RemovePodSandbox \"04e794303c0e25905c56bb11e2aff800b10355fdd500d38327bab315b8dc1c3e\" returns successfully" Apr 24 23:35:07.427641 systemd[1]: run-containerd-runc-k8s.io-f49693d76d271b52c256b698c9428ab82f55700cb8db9e6dfb5c2f74e9220be0-runc.F9I9Jz.mount: Deactivated successfully. Apr 24 23:35:11.499302 kubelet[2738]: I0424 23:35:11.499235 2738 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-kflpl" podStartSLOduration=58.348769893 podStartE2EDuration="1m7.499222502s" podCreationTimestamp="2026-04-24 23:34:04 +0000 UTC" firstStartedPulling="2026-04-24 23:34:31.544384478 +0000 UTC m=+42.516535812" lastFinishedPulling="2026-04-24 23:34:40.694837077 +0000 UTC m=+51.666988421" observedRunningTime="2026-04-24 23:34:41.448793449 +0000 UTC m=+52.420944823" watchObservedRunningTime="2026-04-24 23:35:11.499222502 +0000 UTC m=+82.471373836" Apr 24 23:35:12.313773 systemd[1]: Started sshd@7-95.217.179.164:22-4.175.71.9:33388.service - OpenSSH per-connection server daemon (4.175.71.9:33388). Apr 24 23:35:12.554376 sshd[6270]: Accepted publickey for core from 4.175.71.9 port 33388 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:12.559002 sshd[6270]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:12.575043 systemd-logind[1610]: New session 8 of user core. Apr 24 23:35:12.578734 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 24 23:35:12.840624 sshd[6270]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:12.846986 systemd[1]: sshd@7-95.217.179.164:22-4.175.71.9:33388.service: Deactivated successfully. Apr 24 23:35:12.850599 systemd-logind[1610]: Session 8 logged out. Waiting for processes to exit. Apr 24 23:35:12.850792 systemd[1]: session-8.scope: Deactivated successfully. Apr 24 23:35:12.853101 systemd-logind[1610]: Removed session 8. Apr 24 23:35:17.879035 systemd[1]: Started sshd@8-95.217.179.164:22-4.175.71.9:47614.service - OpenSSH per-connection server daemon (4.175.71.9:47614). Apr 24 23:35:18.118933 sshd[6311]: Accepted publickey for core from 4.175.71.9 port 47614 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:18.123215 sshd[6311]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:18.132912 systemd-logind[1610]: New session 9 of user core. Apr 24 23:35:18.137635 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 24 23:35:18.366859 sshd[6311]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:18.370720 systemd[1]: sshd@8-95.217.179.164:22-4.175.71.9:47614.service: Deactivated successfully. Apr 24 23:35:18.374974 systemd-logind[1610]: Session 9 logged out. Waiting for processes to exit. Apr 24 23:35:18.375333 systemd[1]: session-9.scope: Deactivated successfully. Apr 24 23:35:18.378494 systemd-logind[1610]: Removed session 9. Apr 24 23:35:23.404680 systemd[1]: Started sshd@9-95.217.179.164:22-4.175.71.9:47622.service - OpenSSH per-connection server daemon (4.175.71.9:47622). Apr 24 23:35:23.632522 sshd[6325]: Accepted publickey for core from 4.175.71.9 port 47622 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:23.635186 sshd[6325]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:23.642183 systemd-logind[1610]: New session 10 of user core. Apr 24 23:35:23.657790 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 24 23:35:23.906869 sshd[6325]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:23.912336 systemd[1]: sshd@9-95.217.179.164:22-4.175.71.9:47622.service: Deactivated successfully. Apr 24 23:35:23.922471 systemd[1]: session-10.scope: Deactivated successfully. Apr 24 23:35:23.923567 systemd-logind[1610]: Session 10 logged out. Waiting for processes to exit. Apr 24 23:35:23.927620 systemd-logind[1610]: Removed session 10. Apr 24 23:35:28.946249 systemd[1]: Started sshd@10-95.217.179.164:22-4.175.71.9:43954.service - OpenSSH per-connection server daemon (4.175.71.9:43954). Apr 24 23:35:29.174128 sshd[6342]: Accepted publickey for core from 4.175.71.9 port 43954 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:29.174572 sshd[6342]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:29.185360 systemd-logind[1610]: New session 11 of user core. Apr 24 23:35:29.193650 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 24 23:35:29.434005 sshd[6342]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:29.441057 systemd[1]: sshd@10-95.217.179.164:22-4.175.71.9:43954.service: Deactivated successfully. Apr 24 23:35:29.448825 systemd-logind[1610]: Session 11 logged out. Waiting for processes to exit. Apr 24 23:35:29.449024 systemd[1]: session-11.scope: Deactivated successfully. Apr 24 23:35:29.452217 systemd-logind[1610]: Removed session 11. Apr 24 23:35:34.470848 systemd[1]: Started sshd@11-95.217.179.164:22-4.175.71.9:43970.service - OpenSSH per-connection server daemon (4.175.71.9:43970). Apr 24 23:35:34.677177 sshd[6394]: Accepted publickey for core from 4.175.71.9 port 43970 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:34.679785 sshd[6394]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:34.688822 systemd-logind[1610]: New session 12 of user core. Apr 24 23:35:34.694855 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 24 23:35:34.924775 sshd[6394]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:34.930802 systemd[1]: sshd@11-95.217.179.164:22-4.175.71.9:43970.service: Deactivated successfully. Apr 24 23:35:34.937478 systemd[1]: session-12.scope: Deactivated successfully. Apr 24 23:35:34.939052 systemd-logind[1610]: Session 12 logged out. Waiting for processes to exit. Apr 24 23:35:34.940313 systemd-logind[1610]: Removed session 12. Apr 24 23:35:34.965458 systemd[1]: Started sshd@12-95.217.179.164:22-4.175.71.9:43976.service - OpenSSH per-connection server daemon (4.175.71.9:43976). Apr 24 23:35:35.177593 sshd[6409]: Accepted publickey for core from 4.175.71.9 port 43976 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:35.182537 sshd[6409]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:35.192250 systemd-logind[1610]: New session 13 of user core. Apr 24 23:35:35.202570 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 24 23:35:35.436131 sshd[6409]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:35.439456 systemd[1]: sshd@12-95.217.179.164:22-4.175.71.9:43976.service: Deactivated successfully. Apr 24 23:35:35.442888 systemd-logind[1610]: Session 13 logged out. Waiting for processes to exit. Apr 24 23:35:35.444271 systemd[1]: session-13.scope: Deactivated successfully. Apr 24 23:35:35.445221 systemd-logind[1610]: Removed session 13. Apr 24 23:35:35.473264 systemd[1]: Started sshd@13-95.217.179.164:22-4.175.71.9:48776.service - OpenSSH per-connection server daemon (4.175.71.9:48776). Apr 24 23:35:35.680448 sshd[6422]: Accepted publickey for core from 4.175.71.9 port 48776 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:35.681789 sshd[6422]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:35.690128 systemd-logind[1610]: New session 14 of user core. Apr 24 23:35:35.694791 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 24 23:35:35.939208 sshd[6422]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:35.944830 systemd[1]: sshd@13-95.217.179.164:22-4.175.71.9:48776.service: Deactivated successfully. Apr 24 23:35:35.947428 systemd-logind[1610]: Session 14 logged out. Waiting for processes to exit. Apr 24 23:35:35.947833 systemd[1]: session-14.scope: Deactivated successfully. Apr 24 23:35:35.949678 systemd-logind[1610]: Removed session 14. Apr 24 23:35:40.986126 systemd[1]: Started sshd@14-95.217.179.164:22-4.175.71.9:48792.service - OpenSSH per-connection server daemon (4.175.71.9:48792). Apr 24 23:35:41.203228 sshd[6454]: Accepted publickey for core from 4.175.71.9 port 48792 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:41.206338 sshd[6454]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:41.214692 systemd-logind[1610]: New session 15 of user core. Apr 24 23:35:41.222560 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 24 23:35:41.483522 sshd[6454]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:41.486867 systemd[1]: sshd@14-95.217.179.164:22-4.175.71.9:48792.service: Deactivated successfully. Apr 24 23:35:41.491317 systemd[1]: session-15.scope: Deactivated successfully. Apr 24 23:35:41.491887 systemd-logind[1610]: Session 15 logged out. Waiting for processes to exit. Apr 24 23:35:41.492816 systemd-logind[1610]: Removed session 15. Apr 24 23:35:41.517250 systemd[1]: Started sshd@15-95.217.179.164:22-4.175.71.9:48796.service - OpenSSH per-connection server daemon (4.175.71.9:48796). Apr 24 23:35:41.723417 sshd[6489]: Accepted publickey for core from 4.175.71.9 port 48796 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:41.724695 sshd[6489]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:41.730732 systemd-logind[1610]: New session 16 of user core. Apr 24 23:35:41.738291 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 24 23:35:42.188637 sshd[6489]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:42.193571 systemd[1]: sshd@15-95.217.179.164:22-4.175.71.9:48796.service: Deactivated successfully. Apr 24 23:35:42.202587 systemd[1]: session-16.scope: Deactivated successfully. Apr 24 23:35:42.204447 systemd-logind[1610]: Session 16 logged out. Waiting for processes to exit. Apr 24 23:35:42.207392 systemd-logind[1610]: Removed session 16. Apr 24 23:35:42.226485 systemd[1]: Started sshd@16-95.217.179.164:22-4.175.71.9:48810.service - OpenSSH per-connection server daemon (4.175.71.9:48810). Apr 24 23:35:42.446174 sshd[6500]: Accepted publickey for core from 4.175.71.9 port 48810 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:42.447387 sshd[6500]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:42.452739 systemd-logind[1610]: New session 17 of user core. Apr 24 23:35:42.460621 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 24 23:35:43.099822 sshd[6500]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:43.103191 systemd-logind[1610]: Session 17 logged out. Waiting for processes to exit. Apr 24 23:35:43.104359 systemd[1]: sshd@16-95.217.179.164:22-4.175.71.9:48810.service: Deactivated successfully. Apr 24 23:35:43.107718 systemd[1]: session-17.scope: Deactivated successfully. Apr 24 23:35:43.109772 systemd-logind[1610]: Removed session 17. Apr 24 23:35:43.134422 systemd[1]: Started sshd@17-95.217.179.164:22-4.175.71.9:48824.service - OpenSSH per-connection server daemon (4.175.71.9:48824). Apr 24 23:35:43.347072 sshd[6537]: Accepted publickey for core from 4.175.71.9 port 48824 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:43.350131 sshd[6537]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:43.366303 systemd-logind[1610]: New session 18 of user core. Apr 24 23:35:43.373314 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 24 23:35:43.709586 sshd[6537]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:43.712650 systemd[1]: sshd@17-95.217.179.164:22-4.175.71.9:48824.service: Deactivated successfully. Apr 24 23:35:43.717185 systemd-logind[1610]: Session 18 logged out. Waiting for processes to exit. Apr 24 23:35:43.718180 systemd[1]: session-18.scope: Deactivated successfully. Apr 24 23:35:43.719463 systemd-logind[1610]: Removed session 18. Apr 24 23:35:43.745374 systemd[1]: Started sshd@18-95.217.179.164:22-4.175.71.9:48826.service - OpenSSH per-connection server daemon (4.175.71.9:48826). Apr 24 23:35:43.947252 sshd[6549]: Accepted publickey for core from 4.175.71.9 port 48826 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:43.950307 sshd[6549]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:43.956984 systemd-logind[1610]: New session 19 of user core. Apr 24 23:35:43.965180 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 24 23:35:44.190813 sshd[6549]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:44.197360 systemd-logind[1610]: Session 19 logged out. Waiting for processes to exit. Apr 24 23:35:44.197602 systemd[1]: sshd@18-95.217.179.164:22-4.175.71.9:48826.service: Deactivated successfully. Apr 24 23:35:44.200974 systemd[1]: session-19.scope: Deactivated successfully. Apr 24 23:35:44.202302 systemd-logind[1610]: Removed session 19. Apr 24 23:35:49.235559 systemd[1]: Started sshd@19-95.217.179.164:22-4.175.71.9:60170.service - OpenSSH per-connection server daemon (4.175.71.9:60170). Apr 24 23:35:49.461840 sshd[6567]: Accepted publickey for core from 4.175.71.9 port 60170 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:49.464677 sshd[6567]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:49.472476 systemd-logind[1610]: New session 20 of user core. Apr 24 23:35:49.477588 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 24 23:35:49.717233 sshd[6567]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:49.722154 systemd[1]: sshd@19-95.217.179.164:22-4.175.71.9:60170.service: Deactivated successfully. Apr 24 23:35:49.722431 systemd-logind[1610]: Session 20 logged out. Waiting for processes to exit. Apr 24 23:35:49.724893 systemd[1]: session-20.scope: Deactivated successfully. Apr 24 23:35:49.725851 systemd-logind[1610]: Removed session 20. Apr 24 23:35:54.756971 systemd[1]: Started sshd@20-95.217.179.164:22-4.175.71.9:60178.service - OpenSSH per-connection server daemon (4.175.71.9:60178). Apr 24 23:35:54.991265 sshd[6591]: Accepted publickey for core from 4.175.71.9 port 60178 ssh2: RSA SHA256:/LB5UM8JE+Gm8PLCmanmk+IzzQFWk//dmRsy5hU4ZbM Apr 24 23:35:54.994521 sshd[6591]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 23:35:55.003051 systemd-logind[1610]: New session 21 of user core. Apr 24 23:35:55.010684 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 24 23:35:55.221067 sshd[6591]: pam_unix(sshd:session): session closed for user core Apr 24 23:35:55.224683 systemd[1]: sshd@20-95.217.179.164:22-4.175.71.9:60178.service: Deactivated successfully. Apr 24 23:35:55.228151 systemd-logind[1610]: Session 21 logged out. Waiting for processes to exit. Apr 24 23:35:55.228703 systemd[1]: session-21.scope: Deactivated successfully. Apr 24 23:35:55.231792 systemd-logind[1610]: Removed session 21. Apr 24 23:36:11.464849 systemd[1]: run-containerd-runc-k8s.io-203ba8e79a4d43760a778e6b37541469774f0c2f9a256d4b8cf3c6f2cc3f3156-runc.b6lh5K.mount: Deactivated successfully. Apr 24 23:36:13.053632 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1-rootfs.mount: Deactivated successfully. Apr 24 23:36:13.056395 containerd[1631]: time="2026-04-24T23:36:13.055893163Z" level=info msg="shim disconnected" id=e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1 namespace=k8s.io Apr 24 23:36:13.056395 containerd[1631]: time="2026-04-24T23:36:13.055955912Z" level=warning msg="cleaning up after shim disconnected" id=e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1 namespace=k8s.io Apr 24 23:36:13.056395 containerd[1631]: time="2026-04-24T23:36:13.055964122Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 23:36:13.070663 containerd[1631]: time="2026-04-24T23:36:13.070612845Z" level=warning msg="cleanup warnings time=\"2026-04-24T23:36:13Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Apr 24 23:36:13.073508 containerd[1631]: time="2026-04-24T23:36:13.073046589Z" level=info msg="shim disconnected" id=94229b8631e59bfe3a1dc54f88abcb1f153e15884a228cf13cfb1d456c99f77d namespace=k8s.io Apr 24 23:36:13.073508 containerd[1631]: time="2026-04-24T23:36:13.073136319Z" level=warning msg="cleaning up after shim disconnected" id=94229b8631e59bfe3a1dc54f88abcb1f153e15884a228cf13cfb1d456c99f77d namespace=k8s.io Apr 24 23:36:13.073508 containerd[1631]: time="2026-04-24T23:36:13.073144979Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 23:36:13.082747 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-94229b8631e59bfe3a1dc54f88abcb1f153e15884a228cf13cfb1d456c99f77d-rootfs.mount: Deactivated successfully. Apr 24 23:36:13.084440 containerd[1631]: time="2026-04-24T23:36:13.084269215Z" level=warning msg="cleanup warnings time=\"2026-04-24T23:36:13Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Apr 24 23:36:13.140479 kubelet[2738]: E0424 23:36:13.140438 2738 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:59444->10.0.0.2:2379: read: connection timed out" Apr 24 23:36:13.166834 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c59e36393a4c3113197c00ce89282d2b2c6fedd28206f840df82a8ae54fdd6a7-rootfs.mount: Deactivated successfully. Apr 24 23:36:13.167223 containerd[1631]: time="2026-04-24T23:36:13.167128500Z" level=info msg="shim disconnected" id=c59e36393a4c3113197c00ce89282d2b2c6fedd28206f840df82a8ae54fdd6a7 namespace=k8s.io Apr 24 23:36:13.167223 containerd[1631]: time="2026-04-24T23:36:13.167169680Z" level=warning msg="cleaning up after shim disconnected" id=c59e36393a4c3113197c00ce89282d2b2c6fedd28206f840df82a8ae54fdd6a7 namespace=k8s.io Apr 24 23:36:13.167223 containerd[1631]: time="2026-04-24T23:36:13.167176630Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 23:36:13.669143 kubelet[2738]: I0424 23:36:13.668809 2738 scope.go:117] "RemoveContainer" containerID="c59e36393a4c3113197c00ce89282d2b2c6fedd28206f840df82a8ae54fdd6a7" Apr 24 23:36:13.672692 kubelet[2738]: I0424 23:36:13.672382 2738 scope.go:117] "RemoveContainer" containerID="94229b8631e59bfe3a1dc54f88abcb1f153e15884a228cf13cfb1d456c99f77d" Apr 24 23:36:13.672823 containerd[1631]: time="2026-04-24T23:36:13.672533954Z" level=info msg="CreateContainer within sandbox \"788cfc01e78ddecdfecdaf0a48c7d7cfa593f71c8927e1b3f635ed9199d38df9\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Apr 24 23:36:13.676062 containerd[1631]: time="2026-04-24T23:36:13.675968322Z" level=info msg="CreateContainer within sandbox \"f54fb2fca546265cbee3dcbe31874be2bdba59cf461901e10a9c1711863b52fb\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Apr 24 23:36:13.677670 kubelet[2738]: I0424 23:36:13.677637 2738 scope.go:117] "RemoveContainer" containerID="e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1" Apr 24 23:36:13.680575 containerd[1631]: time="2026-04-24T23:36:13.680375263Z" level=info msg="CreateContainer within sandbox \"4d52a0870a80c56690fea1c02080ca9015dae165b3a40a491b830ba850d7bed3\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Apr 24 23:36:13.705117 containerd[1631]: time="2026-04-24T23:36:13.703999268Z" level=info msg="CreateContainer within sandbox \"f54fb2fca546265cbee3dcbe31874be2bdba59cf461901e10a9c1711863b52fb\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"dfcdbda171a3fcf97e875cc198dd987523862f6c1c20dbf5de5dda40aebeff56\"" Apr 24 23:36:13.709460 containerd[1631]: time="2026-04-24T23:36:13.709427782Z" level=info msg="StartContainer for \"dfcdbda171a3fcf97e875cc198dd987523862f6c1c20dbf5de5dda40aebeff56\"" Apr 24 23:36:13.710822 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1483266803.mount: Deactivated successfully. Apr 24 23:36:13.715949 containerd[1631]: time="2026-04-24T23:36:13.715916119Z" level=info msg="CreateContainer within sandbox \"788cfc01e78ddecdfecdaf0a48c7d7cfa593f71c8927e1b3f635ed9199d38df9\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"9e728536071f645206a7da2bfdc891c9c7ce08836671e0ee13bb21d1995e7606\"" Apr 24 23:36:13.716783 containerd[1631]: time="2026-04-24T23:36:13.716662314Z" level=info msg="StartContainer for \"9e728536071f645206a7da2bfdc891c9c7ce08836671e0ee13bb21d1995e7606\"" Apr 24 23:36:13.720425 containerd[1631]: time="2026-04-24T23:36:13.720285500Z" level=info msg="CreateContainer within sandbox \"4d52a0870a80c56690fea1c02080ca9015dae165b3a40a491b830ba850d7bed3\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"a16ecb2b7c0c81b3c020f70b5fc45b0f483e903c2b9d743f1e446c3ac515dcf8\"" Apr 24 23:36:13.721953 containerd[1631]: time="2026-04-24T23:36:13.720828727Z" level=info msg="StartContainer for \"a16ecb2b7c0c81b3c020f70b5fc45b0f483e903c2b9d743f1e446c3ac515dcf8\"" Apr 24 23:36:13.789069 containerd[1631]: time="2026-04-24T23:36:13.789034157Z" level=info msg="StartContainer for \"dfcdbda171a3fcf97e875cc198dd987523862f6c1c20dbf5de5dda40aebeff56\" returns successfully" Apr 24 23:36:13.806888 containerd[1631]: time="2026-04-24T23:36:13.806858880Z" level=info msg="StartContainer for \"9e728536071f645206a7da2bfdc891c9c7ce08836671e0ee13bb21d1995e7606\" returns successfully" Apr 24 23:36:13.878703 containerd[1631]: time="2026-04-24T23:36:13.878581329Z" level=info msg="StartContainer for \"a16ecb2b7c0c81b3c020f70b5fc45b0f483e903c2b9d743f1e446c3ac515dcf8\" returns successfully" Apr 24 23:36:16.274721 kubelet[2738]: I0424 23:36:16.274596 2738 status_manager.go:895] "Failed to get status for pod" podUID="9e69cda47ba3742cdfd1088ebeda4528" pod="kube-system/kube-apiserver-ci-4081-3-6-n-e88c017743" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:59376->10.0.0.2:2379: read: connection timed out" Apr 24 23:36:17.618586 kubelet[2738]: E0424 23:36:17.618447 2738 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:59284->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4081-3-6-n-e88c017743.18a96f2d08374050 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4081-3-6-n-e88c017743,UID:9e69cda47ba3742cdfd1088ebeda4528,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4081-3-6-n-e88c017743,},FirstTimestamp:2026-04-24 23:36:07.176421456 +0000 UTC m=+138.148572840,LastTimestamp:2026-04-24 23:36:07.176421456 +0000 UTC m=+138.148572840,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-6-n-e88c017743,}" Apr 24 23:36:23.141906 kubelet[2738]: E0424 23:36:23.141579 2738 controller.go:195] "Failed to update lease" err="Put \"https://95.217.179.164:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-e88c017743?timeout=10s\": context deadline exceeded" Apr 24 23:36:25.241271 containerd[1631]: time="2026-04-24T23:36:25.240913899Z" level=info msg="shim disconnected" id=a16ecb2b7c0c81b3c020f70b5fc45b0f483e903c2b9d743f1e446c3ac515dcf8 namespace=k8s.io Apr 24 23:36:25.241271 containerd[1631]: time="2026-04-24T23:36:25.240983318Z" level=warning msg="cleaning up after shim disconnected" id=a16ecb2b7c0c81b3c020f70b5fc45b0f483e903c2b9d743f1e446c3ac515dcf8 namespace=k8s.io Apr 24 23:36:25.241271 containerd[1631]: time="2026-04-24T23:36:25.241011348Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 23:36:25.241783 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a16ecb2b7c0c81b3c020f70b5fc45b0f483e903c2b9d743f1e446c3ac515dcf8-rootfs.mount: Deactivated successfully. Apr 24 23:36:25.719297 kubelet[2738]: I0424 23:36:25.718584 2738 scope.go:117] "RemoveContainer" containerID="e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1" Apr 24 23:36:25.720071 kubelet[2738]: I0424 23:36:25.719564 2738 scope.go:117] "RemoveContainer" containerID="a16ecb2b7c0c81b3c020f70b5fc45b0f483e903c2b9d743f1e446c3ac515dcf8" Apr 24 23:36:25.720071 kubelet[2738]: E0424 23:36:25.719750 2738 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tigera-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=tigera-operator pod=tigera-operator-6bf85f8dd-97vrl_tigera-operator(db1524da-1b6e-4c92-8551-998a8c0db483)\"" pod="tigera-operator/tigera-operator-6bf85f8dd-97vrl" podUID="db1524da-1b6e-4c92-8551-998a8c0db483" Apr 24 23:36:25.722154 containerd[1631]: time="2026-04-24T23:36:25.721898196Z" level=info msg="RemoveContainer for \"e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1\"" Apr 24 23:36:25.728487 containerd[1631]: time="2026-04-24T23:36:25.728430931Z" level=info msg="RemoveContainer for \"e91da3b29dc03eeacc70ea3c6db1f9a2b58242c505600b6cdf94ba27146f15f1\" returns successfully" Apr 24 23:36:29.513324 systemd[1]: run-containerd-runc-k8s.io-b5284ea3a90f10dfe43b73c58f5aab2917b4b9dc0fd4594ef13f0a114b391f77-runc.iIWxud.mount: Deactivated successfully.