Mar 7 01:12:38.991243 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Mar 6 22:58:19 -00 2026 Mar 7 01:12:38.991260 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=531e046a631dbba7b4aae1b7955ffa961f5ce7d570e89a624d767cf739ab70b5 Mar 7 01:12:38.991269 kernel: BIOS-provided physical RAM map: Mar 7 01:12:38.991274 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 7 01:12:38.991278 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ed3efff] usable Mar 7 01:12:38.991282 kernel: BIOS-e820: [mem 0x000000007ed3f000-0x000000007edfffff] reserved Mar 7 01:12:38.991287 kernel: BIOS-e820: [mem 0x000000007ee00000-0x000000007f8ecfff] usable Mar 7 01:12:38.991292 kernel: BIOS-e820: [mem 0x000000007f8ed000-0x000000007f9ecfff] reserved Mar 7 01:12:38.991296 kernel: BIOS-e820: [mem 0x000000007f9ed000-0x000000007faecfff] type 20 Mar 7 01:12:38.991301 kernel: BIOS-e820: [mem 0x000000007faed000-0x000000007fb6cfff] reserved Mar 7 01:12:38.991305 kernel: BIOS-e820: [mem 0x000000007fb6d000-0x000000007fb7efff] ACPI data Mar 7 01:12:38.991312 kernel: BIOS-e820: [mem 0x000000007fb7f000-0x000000007fbfefff] ACPI NVS Mar 7 01:12:38.991316 kernel: BIOS-e820: [mem 0x000000007fbff000-0x000000007ff7bfff] usable Mar 7 01:12:38.991321 kernel: BIOS-e820: [mem 0x000000007ff7c000-0x000000007fffffff] reserved Mar 7 01:12:38.991326 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Mar 7 01:12:38.991331 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 7 01:12:38.991338 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Mar 7 01:12:38.991343 kernel: BIOS-e820: [mem 0x0000000100000000-0x0000000179ffffff] usable Mar 7 01:12:38.991347 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 7 01:12:38.991352 kernel: NX (Execute Disable) protection: active Mar 7 01:12:38.991356 kernel: APIC: Static calls initialized Mar 7 01:12:38.991361 kernel: efi: EFI v2.7 by Ubuntu distribution of EDK II Mar 7 01:12:38.991366 kernel: efi: SMBIOS=0x7f988000 SMBIOS 3.0=0x7f986000 ACPI=0x7fb7e000 ACPI 2.0=0x7fb7e014 MEMATTR=0x7e845198 Mar 7 01:12:38.991371 kernel: efi: Remove mem135: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Mar 7 01:12:38.991375 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Mar 7 01:12:38.991380 kernel: SMBIOS 3.0.0 present. Mar 7 01:12:38.991385 kernel: DMI: Hetzner vServer/Standard PC (Q35 + ICH9, 2009), BIOS 20171111 11/11/2017 Mar 7 01:12:38.991390 kernel: Hypervisor detected: KVM Mar 7 01:12:38.991397 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 7 01:12:38.991401 kernel: kvm-clock: using sched offset of 12649753489 cycles Mar 7 01:12:38.991406 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 7 01:12:38.991411 kernel: tsc: Detected 2399.998 MHz processor Mar 7 01:12:38.991416 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 7 01:12:38.991421 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 7 01:12:38.991426 kernel: last_pfn = 0x17a000 max_arch_pfn = 0x10000000000 Mar 7 01:12:38.991431 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Mar 7 01:12:38.991436 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 7 01:12:38.991442 kernel: last_pfn = 0x7ff7c max_arch_pfn = 0x10000000000 Mar 7 01:12:38.991447 kernel: Using GB pages for direct mapping Mar 7 01:12:38.991452 kernel: Secure boot disabled Mar 7 01:12:38.991861 kernel: ACPI: Early table checksum verification disabled Mar 7 01:12:38.991867 kernel: ACPI: RSDP 0x000000007FB7E014 000024 (v02 BOCHS ) Mar 7 01:12:38.991872 kernel: ACPI: XSDT 0x000000007FB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Mar 7 01:12:38.991877 kernel: ACPI: FACP 0x000000007FB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 01:12:38.991885 kernel: ACPI: DSDT 0x000000007FB7A000 002443 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 01:12:38.991890 kernel: ACPI: FACS 0x000000007FBDD000 000040 Mar 7 01:12:38.991895 kernel: ACPI: APIC 0x000000007FB78000 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 01:12:38.991900 kernel: ACPI: HPET 0x000000007FB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 01:12:38.991905 kernel: ACPI: MCFG 0x000000007FB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 01:12:38.991910 kernel: ACPI: WAET 0x000000007FB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 01:12:38.991915 kernel: ACPI: BGRT 0x000000007FB74000 000038 (v01 INTEL EDK2 00000002 01000013) Mar 7 01:12:38.991923 kernel: ACPI: Reserving FACP table memory at [mem 0x7fb79000-0x7fb790f3] Mar 7 01:12:38.991928 kernel: ACPI: Reserving DSDT table memory at [mem 0x7fb7a000-0x7fb7c442] Mar 7 01:12:38.991933 kernel: ACPI: Reserving FACS table memory at [mem 0x7fbdd000-0x7fbdd03f] Mar 7 01:12:38.991938 kernel: ACPI: Reserving APIC table memory at [mem 0x7fb78000-0x7fb7807f] Mar 7 01:12:38.991943 kernel: ACPI: Reserving HPET table memory at [mem 0x7fb77000-0x7fb77037] Mar 7 01:12:38.991947 kernel: ACPI: Reserving MCFG table memory at [mem 0x7fb76000-0x7fb7603b] Mar 7 01:12:38.991952 kernel: ACPI: Reserving WAET table memory at [mem 0x7fb75000-0x7fb75027] Mar 7 01:12:38.991958 kernel: ACPI: Reserving BGRT table memory at [mem 0x7fb74000-0x7fb74037] Mar 7 01:12:38.991963 kernel: No NUMA configuration found Mar 7 01:12:38.991976 kernel: Faking a node at [mem 0x0000000000000000-0x0000000179ffffff] Mar 7 01:12:38.991983 kernel: NODE_DATA(0) allocated [mem 0x179ff8000-0x179ffdfff] Mar 7 01:12:38.991991 kernel: Zone ranges: Mar 7 01:12:38.991998 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 7 01:12:38.992005 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Mar 7 01:12:38.992013 kernel: Normal [mem 0x0000000100000000-0x0000000179ffffff] Mar 7 01:12:38.992020 kernel: Movable zone start for each node Mar 7 01:12:38.992031 kernel: Early memory node ranges Mar 7 01:12:38.992039 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Mar 7 01:12:38.992046 kernel: node 0: [mem 0x0000000000100000-0x000000007ed3efff] Mar 7 01:12:38.992057 kernel: node 0: [mem 0x000000007ee00000-0x000000007f8ecfff] Mar 7 01:12:38.992064 kernel: node 0: [mem 0x000000007fbff000-0x000000007ff7bfff] Mar 7 01:12:38.992071 kernel: node 0: [mem 0x0000000100000000-0x0000000179ffffff] Mar 7 01:12:38.992077 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x0000000179ffffff] Mar 7 01:12:38.992084 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 7 01:12:38.992092 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Mar 7 01:12:38.992099 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Mar 7 01:12:38.992106 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Mar 7 01:12:38.992112 kernel: On node 0, zone Normal: 132 pages in unavailable ranges Mar 7 01:12:38.992122 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Mar 7 01:12:38.992129 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 7 01:12:38.992136 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 7 01:12:38.992143 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 7 01:12:38.992151 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 7 01:12:38.992158 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 7 01:12:38.992165 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 7 01:12:38.992172 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 7 01:12:38.992179 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 7 01:12:38.992189 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 7 01:12:38.992196 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 7 01:12:38.992204 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Mar 7 01:12:38.992211 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 7 01:12:38.992219 kernel: [mem 0x80000000-0xdfffffff] available for PCI devices Mar 7 01:12:38.992227 kernel: Booting paravirtualized kernel on KVM Mar 7 01:12:38.992235 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 7 01:12:38.992243 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Mar 7 01:12:38.992250 kernel: percpu: Embedded 57 pages/cpu s196328 r8192 d28952 u1048576 Mar 7 01:12:38.992261 kernel: pcpu-alloc: s196328 r8192 d28952 u1048576 alloc=1*2097152 Mar 7 01:12:38.992268 kernel: pcpu-alloc: [0] 0 1 Mar 7 01:12:38.992276 kernel: kvm-guest: PV spinlocks disabled, no host support Mar 7 01:12:38.992284 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=531e046a631dbba7b4aae1b7955ffa961f5ce7d570e89a624d767cf739ab70b5 Mar 7 01:12:38.992290 kernel: random: crng init done Mar 7 01:12:38.992295 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 7 01:12:38.992299 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 7 01:12:38.992304 kernel: Fallback order for Node 0: 0 Mar 7 01:12:38.992312 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1004632 Mar 7 01:12:38.992317 kernel: Policy zone: Normal Mar 7 01:12:38.992322 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 7 01:12:38.992327 kernel: software IO TLB: area num 2. Mar 7 01:12:38.992332 kernel: Memory: 3827828K/4091168K available (12288K kernel code, 2288K rwdata, 22752K rodata, 42892K init, 2304K bss, 263136K reserved, 0K cma-reserved) Mar 7 01:12:38.992337 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 7 01:12:38.992342 kernel: ftrace: allocating 37996 entries in 149 pages Mar 7 01:12:38.992347 kernel: ftrace: allocated 149 pages with 4 groups Mar 7 01:12:38.992351 kernel: Dynamic Preempt: voluntary Mar 7 01:12:38.992359 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 7 01:12:38.992365 kernel: rcu: RCU event tracing is enabled. Mar 7 01:12:38.992370 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 7 01:12:38.992375 kernel: Trampoline variant of Tasks RCU enabled. Mar 7 01:12:38.992387 kernel: Rude variant of Tasks RCU enabled. Mar 7 01:12:38.992395 kernel: Tracing variant of Tasks RCU enabled. Mar 7 01:12:38.992400 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 7 01:12:38.992405 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 7 01:12:38.992410 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Mar 7 01:12:38.992415 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 7 01:12:38.992420 kernel: Console: colour dummy device 80x25 Mar 7 01:12:38.992426 kernel: printk: console [tty0] enabled Mar 7 01:12:38.992433 kernel: printk: console [ttyS0] enabled Mar 7 01:12:38.992439 kernel: ACPI: Core revision 20230628 Mar 7 01:12:38.992444 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 7 01:12:38.992449 kernel: APIC: Switch to symmetric I/O mode setup Mar 7 01:12:38.992454 kernel: x2apic enabled Mar 7 01:12:38.994579 kernel: APIC: Switched APIC routing to: physical x2apic Mar 7 01:12:38.994586 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 7 01:12:38.994592 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Mar 7 01:12:38.994597 kernel: Calibrating delay loop (skipped) preset value.. 4799.99 BogoMIPS (lpj=2399998) Mar 7 01:12:38.994603 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 7 01:12:38.994608 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Mar 7 01:12:38.994613 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Mar 7 01:12:38.994618 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 7 01:12:38.994624 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Mar 7 01:12:38.994632 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Mar 7 01:12:38.994650 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Mar 7 01:12:38.994655 kernel: active return thunk: srso_alias_return_thunk Mar 7 01:12:38.994660 kernel: Speculative Return Stack Overflow: Mitigation: Safe RET Mar 7 01:12:38.994666 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Mar 7 01:12:38.994671 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Mar 7 01:12:38.994676 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 7 01:12:38.994682 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 7 01:12:38.994687 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 7 01:12:38.994694 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Mar 7 01:12:38.994700 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Mar 7 01:12:38.994705 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Mar 7 01:12:38.994711 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Mar 7 01:12:38.994716 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 7 01:12:38.994721 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Mar 7 01:12:38.994727 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Mar 7 01:12:38.994732 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Mar 7 01:12:38.994737 kernel: x86/fpu: xstate_offset[9]: 2432, xstate_sizes[9]: 8 Mar 7 01:12:38.994744 kernel: x86/fpu: Enabled xstate features 0x2e7, context size is 2440 bytes, using 'compacted' format. Mar 7 01:12:38.994750 kernel: Freeing SMP alternatives memory: 32K Mar 7 01:12:38.994755 kernel: pid_max: default: 32768 minimum: 301 Mar 7 01:12:38.994760 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 7 01:12:38.994765 kernel: landlock: Up and running. Mar 7 01:12:38.994771 kernel: SELinux: Initializing. Mar 7 01:12:38.994776 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 7 01:12:38.994781 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 7 01:12:38.994786 kernel: smpboot: CPU0: AMD EPYC-Genoa Processor (family: 0x19, model: 0x11, stepping: 0x0) Mar 7 01:12:38.994794 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 01:12:38.994799 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 01:12:38.994804 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 01:12:38.994809 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Mar 7 01:12:38.994815 kernel: ... version: 0 Mar 7 01:12:38.994820 kernel: ... bit width: 48 Mar 7 01:12:38.994825 kernel: ... generic registers: 6 Mar 7 01:12:38.994830 kernel: ... value mask: 0000ffffffffffff Mar 7 01:12:38.994836 kernel: ... max period: 00007fffffffffff Mar 7 01:12:38.994843 kernel: ... fixed-purpose events: 0 Mar 7 01:12:38.994848 kernel: ... event mask: 000000000000003f Mar 7 01:12:38.994853 kernel: signal: max sigframe size: 3376 Mar 7 01:12:38.994859 kernel: rcu: Hierarchical SRCU implementation. Mar 7 01:12:38.994864 kernel: rcu: Max phase no-delay instances is 400. Mar 7 01:12:38.994870 kernel: smp: Bringing up secondary CPUs ... Mar 7 01:12:38.994875 kernel: smpboot: x86: Booting SMP configuration: Mar 7 01:12:38.994880 kernel: .... node #0, CPUs: #1 Mar 7 01:12:38.994885 kernel: smp: Brought up 1 node, 2 CPUs Mar 7 01:12:38.994893 kernel: smpboot: Max logical packages: 1 Mar 7 01:12:38.994898 kernel: smpboot: Total of 2 processors activated (9599.99 BogoMIPS) Mar 7 01:12:38.994903 kernel: devtmpfs: initialized Mar 7 01:12:38.994909 kernel: x86/mm: Memory block size: 128MB Mar 7 01:12:38.994914 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7fb7f000-0x7fbfefff] (524288 bytes) Mar 7 01:12:38.994919 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 7 01:12:38.994924 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 7 01:12:38.994930 kernel: pinctrl core: initialized pinctrl subsystem Mar 7 01:12:38.994935 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 7 01:12:38.994943 kernel: audit: initializing netlink subsys (disabled) Mar 7 01:12:38.994948 kernel: audit: type=2000 audit(1772845957.574:1): state=initialized audit_enabled=0 res=1 Mar 7 01:12:38.994953 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 7 01:12:38.994958 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 7 01:12:38.994964 kernel: cpuidle: using governor menu Mar 7 01:12:38.994969 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 7 01:12:38.994974 kernel: dca service started, version 1.12.1 Mar 7 01:12:38.994979 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Mar 7 01:12:38.994985 kernel: PCI: Using configuration type 1 for base access Mar 7 01:12:38.994992 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 7 01:12:38.994998 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 7 01:12:38.995003 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 7 01:12:38.995008 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 7 01:12:38.995013 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 7 01:12:38.995019 kernel: ACPI: Added _OSI(Module Device) Mar 7 01:12:38.995024 kernel: ACPI: Added _OSI(Processor Device) Mar 7 01:12:38.995029 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 7 01:12:38.995034 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 7 01:12:38.995042 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Mar 7 01:12:38.995047 kernel: ACPI: Interpreter enabled Mar 7 01:12:38.995052 kernel: ACPI: PM: (supports S0 S5) Mar 7 01:12:38.995057 kernel: ACPI: Using IOAPIC for interrupt routing Mar 7 01:12:38.995062 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 7 01:12:38.995068 kernel: PCI: Using E820 reservations for host bridge windows Mar 7 01:12:38.995073 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 7 01:12:38.995078 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 7 01:12:38.995248 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 7 01:12:38.995399 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Mar 7 01:12:38.995545 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Mar 7 01:12:38.995553 kernel: PCI host bridge to bus 0000:00 Mar 7 01:12:38.995665 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 7 01:12:38.995755 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 7 01:12:38.995844 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 7 01:12:38.995936 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xdfffffff window] Mar 7 01:12:38.996024 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Mar 7 01:12:38.996111 kernel: pci_bus 0000:00: root bus resource [mem 0xc000000000-0xc7ffffffff window] Mar 7 01:12:38.996198 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 7 01:12:38.996309 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Mar 7 01:12:38.996413 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x030000 Mar 7 01:12:38.997248 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80000000-0x807fffff pref] Mar 7 01:12:38.998221 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc060500000-0xc060503fff 64bit pref] Mar 7 01:12:38.998351 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8138a000-0x8138afff] Mar 7 01:12:38.999291 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Mar 7 01:12:38.999429 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Mar 7 01:12:38.999576 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 7 01:12:38.999738 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:38.999879 kernel: pci 0000:00:02.0: reg 0x10: [mem 0x81389000-0x81389fff] Mar 7 01:12:39.000014 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:39.000126 kernel: pci 0000:00:02.1: reg 0x10: [mem 0x81388000-0x81388fff] Mar 7 01:12:39.000254 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:39.000383 kernel: pci 0000:00:02.2: reg 0x10: [mem 0x81387000-0x81387fff] Mar 7 01:12:39.003575 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:39.003727 kernel: pci 0000:00:02.3: reg 0x10: [mem 0x81386000-0x81386fff] Mar 7 01:12:39.003864 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:39.004014 kernel: pci 0000:00:02.4: reg 0x10: [mem 0x81385000-0x81385fff] Mar 7 01:12:39.004151 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:39.004276 kernel: pci 0000:00:02.5: reg 0x10: [mem 0x81384000-0x81384fff] Mar 7 01:12:39.004421 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:39.004614 kernel: pci 0000:00:02.6: reg 0x10: [mem 0x81383000-0x81383fff] Mar 7 01:12:39.004766 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:39.004890 kernel: pci 0000:00:02.7: reg 0x10: [mem 0x81382000-0x81382fff] Mar 7 01:12:39.005040 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Mar 7 01:12:39.005166 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x81381000-0x81381fff] Mar 7 01:12:39.005288 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Mar 7 01:12:39.005395 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 7 01:12:39.006704 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Mar 7 01:12:39.006859 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x6040-0x605f] Mar 7 01:12:39.006970 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0x81380000-0x81380fff] Mar 7 01:12:39.007083 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Mar 7 01:12:39.007198 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6000-0x603f] Mar 7 01:12:39.007318 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Mar 7 01:12:39.007507 kernel: pci 0000:01:00.0: reg 0x14: [mem 0x81200000-0x81200fff] Mar 7 01:12:39.007679 kernel: pci 0000:01:00.0: reg 0x20: [mem 0xc060000000-0xc060003fff 64bit pref] Mar 7 01:12:39.007795 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Mar 7 01:12:39.007901 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Mar 7 01:12:39.008005 kernel: pci 0000:00:02.0: bridge window [mem 0x81200000-0x812fffff] Mar 7 01:12:39.008109 kernel: pci 0000:00:02.0: bridge window [mem 0xc060000000-0xc0600fffff 64bit pref] Mar 7 01:12:39.008226 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Mar 7 01:12:39.008339 kernel: pci 0000:02:00.0: reg 0x10: [mem 0x81100000-0x81103fff 64bit] Mar 7 01:12:39.008443 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Mar 7 01:12:39.012314 kernel: pci 0000:00:02.1: bridge window [mem 0x81100000-0x811fffff] Mar 7 01:12:39.012467 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Mar 7 01:12:39.012580 kernel: pci 0000:03:00.0: reg 0x14: [mem 0x81000000-0x81000fff] Mar 7 01:12:39.012700 kernel: pci 0000:03:00.0: reg 0x20: [mem 0xc060100000-0xc060103fff 64bit pref] Mar 7 01:12:39.012813 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Mar 7 01:12:39.012915 kernel: pci 0000:00:02.2: bridge window [mem 0x81000000-0x810fffff] Mar 7 01:12:39.013017 kernel: pci 0000:00:02.2: bridge window [mem 0xc060100000-0xc0601fffff 64bit pref] Mar 7 01:12:39.013131 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Mar 7 01:12:39.013238 kernel: pci 0000:04:00.0: reg 0x20: [mem 0xc060200000-0xc060203fff 64bit pref] Mar 7 01:12:39.013366 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Mar 7 01:12:39.013505 kernel: pci 0000:00:02.3: bridge window [mem 0xc060200000-0xc0602fffff 64bit pref] Mar 7 01:12:39.013619 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Mar 7 01:12:39.013741 kernel: pci 0000:05:00.0: reg 0x14: [mem 0x80f00000-0x80f00fff] Mar 7 01:12:39.013846 kernel: pci 0000:05:00.0: reg 0x20: [mem 0xc060300000-0xc060303fff 64bit pref] Mar 7 01:12:39.013946 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Mar 7 01:12:39.014045 kernel: pci 0000:00:02.4: bridge window [mem 0x80f00000-0x80ffffff] Mar 7 01:12:39.014143 kernel: pci 0000:00:02.4: bridge window [mem 0xc060300000-0xc0603fffff 64bit pref] Mar 7 01:12:39.014255 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Mar 7 01:12:39.014359 kernel: pci 0000:06:00.0: reg 0x14: [mem 0x80e00000-0x80e00fff] Mar 7 01:12:39.014478 kernel: pci 0000:06:00.0: reg 0x20: [mem 0xc060400000-0xc060403fff 64bit pref] Mar 7 01:12:39.014580 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Mar 7 01:12:39.014689 kernel: pci 0000:00:02.5: bridge window [mem 0x80e00000-0x80efffff] Mar 7 01:12:39.014787 kernel: pci 0000:00:02.5: bridge window [mem 0xc060400000-0xc0604fffff 64bit pref] Mar 7 01:12:39.014794 kernel: acpiphp: Slot [0] registered Mar 7 01:12:39.014915 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Mar 7 01:12:39.015027 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x80c00000-0x80c00fff] Mar 7 01:12:39.015135 kernel: pci 0000:07:00.0: reg 0x20: [mem 0xc000000000-0xc000003fff 64bit pref] Mar 7 01:12:39.015238 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Mar 7 01:12:39.015337 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Mar 7 01:12:39.015434 kernel: pci 0000:00:02.6: bridge window [mem 0x80c00000-0x80dfffff] Mar 7 01:12:39.016487 kernel: pci 0000:00:02.6: bridge window [mem 0xc000000000-0xc01fffffff 64bit pref] Mar 7 01:12:39.016501 kernel: acpiphp: Slot [0-2] registered Mar 7 01:12:39.016606 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Mar 7 01:12:39.016714 kernel: pci 0000:00:02.7: bridge window [mem 0x80a00000-0x80bfffff] Mar 7 01:12:39.016812 kernel: pci 0000:00:02.7: bridge window [mem 0xc020000000-0xc03fffffff 64bit pref] Mar 7 01:12:39.016822 kernel: acpiphp: Slot [0-3] registered Mar 7 01:12:39.016921 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Mar 7 01:12:39.017015 kernel: pci 0000:00:03.0: bridge window [mem 0x80800000-0x809fffff] Mar 7 01:12:39.017109 kernel: pci 0000:00:03.0: bridge window [mem 0xc040000000-0xc05fffffff 64bit pref] Mar 7 01:12:39.017115 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 7 01:12:39.017121 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 7 01:12:39.017126 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 7 01:12:39.017132 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 7 01:12:39.017140 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 7 01:12:39.017146 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 7 01:12:39.017151 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 7 01:12:39.017157 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 7 01:12:39.017162 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 7 01:12:39.017167 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 7 01:12:39.017173 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 7 01:12:39.017178 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 7 01:12:39.017184 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 7 01:12:39.017191 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 7 01:12:39.017197 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 7 01:12:39.017202 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 7 01:12:39.017207 kernel: iommu: Default domain type: Translated Mar 7 01:12:39.017213 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 7 01:12:39.017218 kernel: efivars: Registered efivars operations Mar 7 01:12:39.017223 kernel: PCI: Using ACPI for IRQ routing Mar 7 01:12:39.017229 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 7 01:12:39.017235 kernel: e820: reserve RAM buffer [mem 0x7ed3f000-0x7fffffff] Mar 7 01:12:39.017243 kernel: e820: reserve RAM buffer [mem 0x7f8ed000-0x7fffffff] Mar 7 01:12:39.017248 kernel: e820: reserve RAM buffer [mem 0x7ff7c000-0x7fffffff] Mar 7 01:12:39.017253 kernel: e820: reserve RAM buffer [mem 0x17a000000-0x17bffffff] Mar 7 01:12:39.017351 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 7 01:12:39.017446 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 7 01:12:39.017599 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 7 01:12:39.017607 kernel: vgaarb: loaded Mar 7 01:12:39.017613 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 7 01:12:39.017618 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 7 01:12:39.017627 kernel: clocksource: Switched to clocksource kvm-clock Mar 7 01:12:39.017632 kernel: VFS: Disk quotas dquot_6.6.0 Mar 7 01:12:39.017648 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 7 01:12:39.017654 kernel: pnp: PnP ACPI init Mar 7 01:12:39.017763 kernel: system 00:04: [mem 0xe0000000-0xefffffff window] has been reserved Mar 7 01:12:39.017771 kernel: pnp: PnP ACPI: found 5 devices Mar 7 01:12:39.017776 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 7 01:12:39.017782 kernel: NET: Registered PF_INET protocol family Mar 7 01:12:39.017803 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 7 01:12:39.017811 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 7 01:12:39.017817 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 7 01:12:39.017823 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 7 01:12:39.017828 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 7 01:12:39.017834 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 7 01:12:39.017840 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 7 01:12:39.017845 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 7 01:12:39.017851 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 7 01:12:39.017859 kernel: NET: Registered PF_XDP protocol family Mar 7 01:12:39.017962 kernel: pci 0000:01:00.0: can't claim BAR 6 [mem 0xfff80000-0xffffffff pref]: no compatible bridge window Mar 7 01:12:39.018065 kernel: pci 0000:07:00.0: can't claim BAR 6 [mem 0xfff80000-0xffffffff pref]: no compatible bridge window Mar 7 01:12:39.018162 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Mar 7 01:12:39.018260 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Mar 7 01:12:39.018357 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Mar 7 01:12:39.018454 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x1000-0x1fff] Mar 7 01:12:39.018565 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x2000-0x2fff] Mar 7 01:12:39.018674 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x3000-0x3fff] Mar 7 01:12:39.018775 kernel: pci 0000:01:00.0: BAR 6: assigned [mem 0x81280000-0x812fffff pref] Mar 7 01:12:39.018872 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Mar 7 01:12:39.018971 kernel: pci 0000:00:02.0: bridge window [mem 0x81200000-0x812fffff] Mar 7 01:12:39.019067 kernel: pci 0000:00:02.0: bridge window [mem 0xc060000000-0xc0600fffff 64bit pref] Mar 7 01:12:39.019164 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Mar 7 01:12:39.019261 kernel: pci 0000:00:02.1: bridge window [mem 0x81100000-0x811fffff] Mar 7 01:12:39.019359 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Mar 7 01:12:39.019560 kernel: pci 0000:00:02.2: bridge window [mem 0x81000000-0x810fffff] Mar 7 01:12:39.019687 kernel: pci 0000:00:02.2: bridge window [mem 0xc060100000-0xc0601fffff 64bit pref] Mar 7 01:12:39.019797 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Mar 7 01:12:39.019893 kernel: pci 0000:00:02.3: bridge window [mem 0xc060200000-0xc0602fffff 64bit pref] Mar 7 01:12:39.019992 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Mar 7 01:12:39.020087 kernel: pci 0000:00:02.4: bridge window [mem 0x80f00000-0x80ffffff] Mar 7 01:12:39.020181 kernel: pci 0000:00:02.4: bridge window [mem 0xc060300000-0xc0603fffff 64bit pref] Mar 7 01:12:39.020279 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Mar 7 01:12:39.020373 kernel: pci 0000:00:02.5: bridge window [mem 0x80e00000-0x80efffff] Mar 7 01:12:39.020479 kernel: pci 0000:00:02.5: bridge window [mem 0xc060400000-0xc0604fffff 64bit pref] Mar 7 01:12:39.020591 kernel: pci 0000:07:00.0: BAR 6: assigned [mem 0x80c80000-0x80cfffff pref] Mar 7 01:12:39.020701 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Mar 7 01:12:39.020796 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff] Mar 7 01:12:39.020890 kernel: pci 0000:00:02.6: bridge window [mem 0x80c00000-0x80dfffff] Mar 7 01:12:39.020987 kernel: pci 0000:00:02.6: bridge window [mem 0xc000000000-0xc01fffffff 64bit pref] Mar 7 01:12:39.021083 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Mar 7 01:12:39.021179 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff] Mar 7 01:12:39.021273 kernel: pci 0000:00:02.7: bridge window [mem 0x80a00000-0x80bfffff] Mar 7 01:12:39.021368 kernel: pci 0000:00:02.7: bridge window [mem 0xc020000000-0xc03fffffff 64bit pref] Mar 7 01:12:39.023487 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Mar 7 01:12:39.023604 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff] Mar 7 01:12:39.023717 kernel: pci 0000:00:03.0: bridge window [mem 0x80800000-0x809fffff] Mar 7 01:12:39.023815 kernel: pci 0000:00:03.0: bridge window [mem 0xc040000000-0xc05fffffff 64bit pref] Mar 7 01:12:39.023910 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 7 01:12:39.024000 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 7 01:12:39.024091 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 7 01:12:39.024179 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xdfffffff window] Mar 7 01:12:39.024266 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Mar 7 01:12:39.024353 kernel: pci_bus 0000:00: resource 9 [mem 0xc000000000-0xc7ffffffff window] Mar 7 01:12:39.024454 kernel: pci_bus 0000:01: resource 1 [mem 0x81200000-0x812fffff] Mar 7 01:12:39.026593 kernel: pci_bus 0000:01: resource 2 [mem 0xc060000000-0xc0600fffff 64bit pref] Mar 7 01:12:39.026713 kernel: pci_bus 0000:02: resource 1 [mem 0x81100000-0x811fffff] Mar 7 01:12:39.026823 kernel: pci_bus 0000:03: resource 1 [mem 0x81000000-0x810fffff] Mar 7 01:12:39.026916 kernel: pci_bus 0000:03: resource 2 [mem 0xc060100000-0xc0601fffff 64bit pref] Mar 7 01:12:39.027014 kernel: pci_bus 0000:04: resource 2 [mem 0xc060200000-0xc0602fffff 64bit pref] Mar 7 01:12:39.027114 kernel: pci_bus 0000:05: resource 1 [mem 0x80f00000-0x80ffffff] Mar 7 01:12:39.027206 kernel: pci_bus 0000:05: resource 2 [mem 0xc060300000-0xc0603fffff 64bit pref] Mar 7 01:12:39.027304 kernel: pci_bus 0000:06: resource 1 [mem 0x80e00000-0x80efffff] Mar 7 01:12:39.027400 kernel: pci_bus 0000:06: resource 2 [mem 0xc060400000-0xc0604fffff 64bit pref] Mar 7 01:12:39.027570 kernel: pci_bus 0000:07: resource 0 [io 0x1000-0x1fff] Mar 7 01:12:39.027673 kernel: pci_bus 0000:07: resource 1 [mem 0x80c00000-0x80dfffff] Mar 7 01:12:39.027764 kernel: pci_bus 0000:07: resource 2 [mem 0xc000000000-0xc01fffffff 64bit pref] Mar 7 01:12:39.027862 kernel: pci_bus 0000:08: resource 0 [io 0x2000-0x2fff] Mar 7 01:12:39.027954 kernel: pci_bus 0000:08: resource 1 [mem 0x80a00000-0x80bfffff] Mar 7 01:12:39.028045 kernel: pci_bus 0000:08: resource 2 [mem 0xc020000000-0xc03fffffff 64bit pref] Mar 7 01:12:39.028145 kernel: pci_bus 0000:09: resource 0 [io 0x3000-0x3fff] Mar 7 01:12:39.028237 kernel: pci_bus 0000:09: resource 1 [mem 0x80800000-0x809fffff] Mar 7 01:12:39.028328 kernel: pci_bus 0000:09: resource 2 [mem 0xc040000000-0xc05fffffff 64bit pref] Mar 7 01:12:39.028335 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 7 01:12:39.028342 kernel: PCI: CLS 0 bytes, default 64 Mar 7 01:12:39.028347 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Mar 7 01:12:39.028353 kernel: software IO TLB: mapped [mem 0x0000000077ffd000-0x000000007bffd000] (64MB) Mar 7 01:12:39.028361 kernel: Initialise system trusted keyrings Mar 7 01:12:39.028367 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 7 01:12:39.028373 kernel: Key type asymmetric registered Mar 7 01:12:39.028378 kernel: Asymmetric key parser 'x509' registered Mar 7 01:12:39.028384 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Mar 7 01:12:39.028390 kernel: io scheduler mq-deadline registered Mar 7 01:12:39.028395 kernel: io scheduler kyber registered Mar 7 01:12:39.028401 kernel: io scheduler bfq registered Mar 7 01:12:39.028513 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Mar 7 01:12:39.028612 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Mar 7 01:12:39.028721 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Mar 7 01:12:39.028817 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Mar 7 01:12:39.028913 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Mar 7 01:12:39.029007 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Mar 7 01:12:39.029106 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Mar 7 01:12:39.029202 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Mar 7 01:12:39.029298 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Mar 7 01:12:39.029394 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Mar 7 01:12:39.030211 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Mar 7 01:12:39.030317 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Mar 7 01:12:39.030416 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Mar 7 01:12:39.030593 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Mar 7 01:12:39.030702 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Mar 7 01:12:39.030798 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Mar 7 01:12:39.030805 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 7 01:12:39.030899 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 32 Mar 7 01:12:39.030997 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 32 Mar 7 01:12:39.031003 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 7 01:12:39.031010 kernel: ACPI: \_SB_.GSIF: Enabled at IRQ 21 Mar 7 01:12:39.031015 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 7 01:12:39.031021 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 7 01:12:39.031027 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 7 01:12:39.031033 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 7 01:12:39.031038 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 7 01:12:39.031141 kernel: rtc_cmos 00:03: RTC can wake from S4 Mar 7 01:12:39.031152 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Mar 7 01:12:39.031241 kernel: rtc_cmos 00:03: registered as rtc0 Mar 7 01:12:39.031331 kernel: rtc_cmos 00:03: setting system clock to 2026-03-07T01:12:38 UTC (1772845958) Mar 7 01:12:39.031420 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Mar 7 01:12:39.031426 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Mar 7 01:12:39.031433 kernel: efifb: probing for efifb Mar 7 01:12:39.031438 kernel: efifb: framebuffer at 0x80000000, using 4032k, total 4032k Mar 7 01:12:39.031446 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Mar 7 01:12:39.031452 kernel: efifb: scrolling: redraw Mar 7 01:12:39.031468 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 7 01:12:39.031474 kernel: Console: switching to colour frame buffer device 160x50 Mar 7 01:12:39.031479 kernel: fb0: EFI VGA frame buffer device Mar 7 01:12:39.031485 kernel: pstore: Using crash dump compression: deflate Mar 7 01:12:39.031490 kernel: pstore: Registered efi_pstore as persistent store backend Mar 7 01:12:39.031496 kernel: NET: Registered PF_INET6 protocol family Mar 7 01:12:39.031502 kernel: Segment Routing with IPv6 Mar 7 01:12:39.031507 kernel: In-situ OAM (IOAM) with IPv6 Mar 7 01:12:39.031516 kernel: NET: Registered PF_PACKET protocol family Mar 7 01:12:39.031521 kernel: Key type dns_resolver registered Mar 7 01:12:39.031527 kernel: IPI shorthand broadcast: enabled Mar 7 01:12:39.031533 kernel: sched_clock: Marking stable (1318011826, 226035457)->(1598134745, -54087462) Mar 7 01:12:39.031538 kernel: registered taskstats version 1 Mar 7 01:12:39.031544 kernel: Loading compiled-in X.509 certificates Mar 7 01:12:39.031550 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: da286e6f6c247ee6f65a875c513de7da57782e90' Mar 7 01:12:39.031555 kernel: Key type .fscrypt registered Mar 7 01:12:39.031561 kernel: Key type fscrypt-provisioning registered Mar 7 01:12:39.031569 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 7 01:12:39.031574 kernel: ima: Allocated hash algorithm: sha1 Mar 7 01:12:39.031580 kernel: ima: No architecture policies found Mar 7 01:12:39.031585 kernel: clk: Disabling unused clocks Mar 7 01:12:39.031591 kernel: Freeing unused kernel image (initmem) memory: 42892K Mar 7 01:12:39.031597 kernel: Write protecting the kernel read-only data: 36864k Mar 7 01:12:39.031602 kernel: Freeing unused kernel image (rodata/data gap) memory: 1824K Mar 7 01:12:39.031608 kernel: Run /init as init process Mar 7 01:12:39.031615 kernel: with arguments: Mar 7 01:12:39.031621 kernel: /init Mar 7 01:12:39.031627 kernel: with environment: Mar 7 01:12:39.031632 kernel: HOME=/ Mar 7 01:12:39.031646 kernel: TERM=linux Mar 7 01:12:39.031653 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 7 01:12:39.031662 systemd[1]: Detected virtualization kvm. Mar 7 01:12:39.031668 systemd[1]: Detected architecture x86-64. Mar 7 01:12:39.031676 systemd[1]: Running in initrd. Mar 7 01:12:39.031682 systemd[1]: No hostname configured, using default hostname. Mar 7 01:12:39.031688 systemd[1]: Hostname set to . Mar 7 01:12:39.031694 systemd[1]: Initializing machine ID from VM UUID. Mar 7 01:12:39.031702 systemd[1]: Queued start job for default target initrd.target. Mar 7 01:12:39.031708 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 01:12:39.031714 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 01:12:39.031721 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 7 01:12:39.031729 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 7 01:12:39.031735 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 7 01:12:39.031741 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 7 01:12:39.031748 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 7 01:12:39.031754 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 7 01:12:39.031760 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 01:12:39.031768 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 7 01:12:39.031774 systemd[1]: Reached target paths.target - Path Units. Mar 7 01:12:39.031780 systemd[1]: Reached target slices.target - Slice Units. Mar 7 01:12:39.031785 systemd[1]: Reached target swap.target - Swaps. Mar 7 01:12:39.031792 systemd[1]: Reached target timers.target - Timer Units. Mar 7 01:12:39.031797 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 7 01:12:39.031803 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 7 01:12:39.031809 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 7 01:12:39.031815 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 7 01:12:39.031824 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 7 01:12:39.031829 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 7 01:12:39.031835 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 01:12:39.031841 systemd[1]: Reached target sockets.target - Socket Units. Mar 7 01:12:39.031847 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 7 01:12:39.031853 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 7 01:12:39.031859 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 7 01:12:39.031865 systemd[1]: Starting systemd-fsck-usr.service... Mar 7 01:12:39.031871 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 7 01:12:39.031879 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 7 01:12:39.031885 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:12:39.031891 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 7 01:12:39.031915 systemd-journald[189]: Collecting audit messages is disabled. Mar 7 01:12:39.031932 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 01:12:39.031938 systemd[1]: Finished systemd-fsck-usr.service. Mar 7 01:12:39.031945 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 7 01:12:39.031951 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 7 01:12:39.031959 kernel: Bridge firewalling registered Mar 7 01:12:39.031965 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 7 01:12:39.031971 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:12:39.031977 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:12:39.031984 systemd-journald[189]: Journal started Mar 7 01:12:39.031997 systemd-journald[189]: Runtime Journal (/run/log/journal/39ce0ed0a9ee49bfabcd2a5641db3a8b) is 8.0M, max 76.3M, 68.3M free. Mar 7 01:12:38.974914 systemd-modules-load[190]: Inserted module 'overlay' Mar 7 01:12:39.014963 systemd-modules-load[190]: Inserted module 'br_netfilter' Mar 7 01:12:39.046475 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 7 01:12:39.049279 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 7 01:12:39.049796 systemd[1]: Started systemd-journald.service - Journal Service. Mar 7 01:12:39.052855 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 7 01:12:39.061579 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 7 01:12:39.062883 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 7 01:12:39.065116 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:12:39.077589 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 7 01:12:39.078126 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 01:12:39.087126 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 01:12:39.090832 dracut-cmdline[219]: dracut-dracut-053 Mar 7 01:12:39.093564 dracut-cmdline[219]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=531e046a631dbba7b4aae1b7955ffa961f5ce7d570e89a624d767cf739ab70b5 Mar 7 01:12:39.096588 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 7 01:12:39.119622 systemd-resolved[230]: Positive Trust Anchors: Mar 7 01:12:39.119663 systemd-resolved[230]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 7 01:12:39.119685 systemd-resolved[230]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 7 01:12:39.123651 systemd-resolved[230]: Defaulting to hostname 'linux'. Mar 7 01:12:39.124594 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 7 01:12:39.125432 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 7 01:12:39.164496 kernel: SCSI subsystem initialized Mar 7 01:12:39.171476 kernel: Loading iSCSI transport class v2.0-870. Mar 7 01:12:39.181589 kernel: iscsi: registered transport (tcp) Mar 7 01:12:39.197730 kernel: iscsi: registered transport (qla4xxx) Mar 7 01:12:39.197771 kernel: QLogic iSCSI HBA Driver Mar 7 01:12:39.240273 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 7 01:12:39.244581 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 7 01:12:39.266560 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 7 01:12:39.266605 kernel: device-mapper: uevent: version 1.0.3 Mar 7 01:12:39.270202 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 7 01:12:39.308493 kernel: raid6: avx512x4 gen() 41983 MB/s Mar 7 01:12:39.326530 kernel: raid6: avx512x2 gen() 44255 MB/s Mar 7 01:12:39.344527 kernel: raid6: avx512x1 gen() 42042 MB/s Mar 7 01:12:39.362506 kernel: raid6: avx2x4 gen() 44228 MB/s Mar 7 01:12:39.380506 kernel: raid6: avx2x2 gen() 47458 MB/s Mar 7 01:12:39.399504 kernel: raid6: avx2x1 gen() 38506 MB/s Mar 7 01:12:39.399571 kernel: raid6: using algorithm avx2x2 gen() 47458 MB/s Mar 7 01:12:39.419559 kernel: raid6: .... xor() 37285 MB/s, rmw enabled Mar 7 01:12:39.419625 kernel: raid6: using avx512x2 recovery algorithm Mar 7 01:12:39.436565 kernel: xor: automatically using best checksumming function avx Mar 7 01:12:39.570514 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 7 01:12:39.587342 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 7 01:12:39.593646 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 01:12:39.633021 systemd-udevd[406]: Using default interface naming scheme 'v255'. Mar 7 01:12:39.638577 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 01:12:39.647582 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 7 01:12:39.659543 dracut-pre-trigger[414]: rd.md=0: removing MD RAID activation Mar 7 01:12:39.688447 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 7 01:12:39.694600 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 7 01:12:39.763813 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 01:12:39.776605 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 7 01:12:39.790019 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 7 01:12:39.792066 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 7 01:12:39.793678 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 01:12:39.794905 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 7 01:12:39.802905 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 7 01:12:39.814215 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 7 01:12:39.854485 kernel: scsi host0: Virtio SCSI HBA Mar 7 01:12:39.859699 kernel: ACPI: bus type USB registered Mar 7 01:12:39.859746 kernel: usbcore: registered new interface driver usbfs Mar 7 01:12:39.861871 kernel: usbcore: registered new interface driver hub Mar 7 01:12:39.861892 kernel: cryptd: max_cpu_qlen set to 1000 Mar 7 01:12:39.861901 kernel: usbcore: registered new device driver usb Mar 7 01:12:39.868673 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Mar 7 01:12:39.885216 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 01:12:39.885750 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:12:39.886679 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:12:39.886986 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:12:39.887095 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:12:39.887425 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:12:39.894796 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:12:39.915687 kernel: AVX2 version of gcm_enc/dec engaged. Mar 7 01:12:39.915738 kernel: AES CTR mode by8 optimization enabled Mar 7 01:12:39.916491 kernel: libata version 3.00 loaded. Mar 7 01:12:39.933293 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:12:39.941988 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Mar 7 01:12:39.942198 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Mar 7 01:12:39.944835 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Mar 7 01:12:39.943575 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:12:39.946597 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Mar 7 01:12:39.948587 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Mar 7 01:12:39.954477 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Mar 7 01:12:39.959632 kernel: ahci 0000:00:1f.2: version 3.0 Mar 7 01:12:39.959796 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 7 01:12:39.959806 kernel: hub 1-0:1.0: USB hub found Mar 7 01:12:39.959951 kernel: hub 1-0:1.0: 4 ports detected Mar 7 01:12:39.966496 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Mar 7 01:12:39.966654 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 7 01:12:39.966769 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Mar 7 01:12:39.973881 kernel: hub 2-0:1.0: USB hub found Mar 7 01:12:39.975396 kernel: hub 2-0:1.0: 4 ports detected Mar 7 01:12:39.986385 kernel: sd 0:0:0:0: Power-on or device reset occurred Mar 7 01:12:39.989942 kernel: sd 0:0:0:0: [sda] 160006144 512-byte logical blocks: (81.9 GB/76.3 GiB) Mar 7 01:12:39.990149 kernel: scsi host1: ahci Mar 7 01:12:39.989595 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:12:39.992985 kernel: sd 0:0:0:0: [sda] Write Protect is off Mar 7 01:12:39.993135 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Mar 7 01:12:39.996094 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Mar 7 01:12:39.996263 kernel: scsi host2: ahci Mar 7 01:12:39.999635 kernel: scsi host3: ahci Mar 7 01:12:39.999835 kernel: scsi host4: ahci Mar 7 01:12:39.999957 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 7 01:12:40.002582 kernel: scsi host5: ahci Mar 7 01:12:40.002614 kernel: GPT:17805311 != 160006143 Mar 7 01:12:40.006474 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 7 01:12:40.006507 kernel: scsi host6: ahci Mar 7 01:12:40.006678 kernel: GPT:17805311 != 160006143 Mar 7 01:12:40.006687 kernel: ata1: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380100 irq 51 Mar 7 01:12:40.006696 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 7 01:12:40.006703 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 01:12:40.006711 kernel: ata2: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380180 irq 51 Mar 7 01:12:40.006719 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Mar 7 01:12:40.006850 kernel: ata3: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380200 irq 51 Mar 7 01:12:40.025364 kernel: ata4: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380280 irq 51 Mar 7 01:12:40.025395 kernel: ata5: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380300 irq 51 Mar 7 01:12:40.025404 kernel: ata6: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380380 irq 51 Mar 7 01:12:40.040312 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Mar 7 01:12:40.043049 kernel: BTRFS: device fsid 3bed8db9-42ad-4483-9cc8-1ad17a6cd948 devid 1 transid 34 /dev/sda3 scanned by (udev-worker) (474) Mar 7 01:12:40.048496 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/sda6 scanned by (udev-worker) (455) Mar 7 01:12:40.047811 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Mar 7 01:12:40.057769 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Mar 7 01:12:40.058884 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Mar 7 01:12:40.062968 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Mar 7 01:12:40.073579 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 7 01:12:40.078766 disk-uuid[580]: Primary Header is updated. Mar 7 01:12:40.078766 disk-uuid[580]: Secondary Entries is updated. Mar 7 01:12:40.078766 disk-uuid[580]: Secondary Header is updated. Mar 7 01:12:40.084223 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 01:12:40.088486 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 01:12:40.204518 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Mar 7 01:12:40.342502 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 7 01:12:40.342583 kernel: ata3: SATA link down (SStatus 0 SControl 300) Mar 7 01:12:40.349911 kernel: ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Mar 7 01:12:40.349955 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 7 01:12:40.363047 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 7 01:12:40.363167 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 7 01:12:40.368515 kernel: ata1.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Mar 7 01:12:40.368559 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 7 01:12:40.368577 kernel: ata1.00: applying bridge limits Mar 7 01:12:40.378144 kernel: ata1.00: configured for UDMA/100 Mar 7 01:12:40.384621 kernel: scsi 1:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 7 01:12:40.425932 kernel: usbcore: registered new interface driver usbhid Mar 7 01:12:40.425996 kernel: usbhid: USB HID core driver Mar 7 01:12:40.438862 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input3 Mar 7 01:12:40.438903 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Mar 7 01:12:40.453536 kernel: sr 1:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Mar 7 01:12:40.453917 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 7 01:12:40.469755 kernel: sr 1:0:0:0: Attached scsi CD-ROM sr0 Mar 7 01:12:41.097504 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 01:12:41.099578 disk-uuid[581]: The operation has completed successfully. Mar 7 01:12:41.175535 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 7 01:12:41.175641 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 7 01:12:41.178638 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 7 01:12:41.184539 sh[599]: Success Mar 7 01:12:41.199581 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Mar 7 01:12:41.237831 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 7 01:12:41.244550 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 7 01:12:41.246667 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 7 01:12:41.263793 kernel: BTRFS info (device dm-0): first mount of filesystem 3bed8db9-42ad-4483-9cc8-1ad17a6cd948 Mar 7 01:12:41.263831 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 7 01:12:41.263840 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 7 01:12:41.266654 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 7 01:12:41.268583 kernel: BTRFS info (device dm-0): using free space tree Mar 7 01:12:41.278491 kernel: BTRFS info (device dm-0): enabling ssd optimizations Mar 7 01:12:41.279726 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 7 01:12:41.280594 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 7 01:12:41.289585 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 7 01:12:41.292783 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 7 01:12:41.309295 kernel: BTRFS info (device sda6): first mount of filesystem 872bf425-12c9-4ef2-aaf0-71379b3513d9 Mar 7 01:12:41.309332 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 7 01:12:41.309342 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:12:41.316010 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 7 01:12:41.316037 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:12:41.329310 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 7 01:12:41.331789 kernel: BTRFS info (device sda6): last unmount of filesystem 872bf425-12c9-4ef2-aaf0-71379b3513d9 Mar 7 01:12:41.338620 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 7 01:12:41.344635 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 7 01:12:41.393684 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 7 01:12:41.402870 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 7 01:12:41.413515 ignition[720]: Ignition 2.19.0 Mar 7 01:12:41.413527 ignition[720]: Stage: fetch-offline Mar 7 01:12:41.413563 ignition[720]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:12:41.413573 ignition[720]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 01:12:41.415711 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 7 01:12:41.413652 ignition[720]: parsed url from cmdline: "" Mar 7 01:12:41.413655 ignition[720]: no config URL provided Mar 7 01:12:41.413660 ignition[720]: reading system config file "/usr/lib/ignition/user.ign" Mar 7 01:12:41.413668 ignition[720]: no config at "/usr/lib/ignition/user.ign" Mar 7 01:12:41.413673 ignition[720]: failed to fetch config: resource requires networking Mar 7 01:12:41.413903 ignition[720]: Ignition finished successfully Mar 7 01:12:41.431952 systemd-networkd[780]: lo: Link UP Mar 7 01:12:41.432419 systemd-networkd[780]: lo: Gained carrier Mar 7 01:12:41.434790 systemd-networkd[780]: Enumeration completed Mar 7 01:12:41.434888 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 7 01:12:41.435441 systemd[1]: Reached target network.target - Network. Mar 7 01:12:41.435772 systemd-networkd[780]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:41.435776 systemd-networkd[780]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 01:12:41.437937 systemd-networkd[780]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:41.437941 systemd-networkd[780]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 01:12:41.439758 systemd-networkd[780]: eth0: Link UP Mar 7 01:12:41.439763 systemd-networkd[780]: eth0: Gained carrier Mar 7 01:12:41.439771 systemd-networkd[780]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:41.444602 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 7 01:12:41.444695 systemd-networkd[780]: eth1: Link UP Mar 7 01:12:41.444699 systemd-networkd[780]: eth1: Gained carrier Mar 7 01:12:41.444706 systemd-networkd[780]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:41.457053 ignition[787]: Ignition 2.19.0 Mar 7 01:12:41.457064 ignition[787]: Stage: fetch Mar 7 01:12:41.457315 ignition[787]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:12:41.457326 ignition[787]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 01:12:41.457424 ignition[787]: parsed url from cmdline: "" Mar 7 01:12:41.457430 ignition[787]: no config URL provided Mar 7 01:12:41.457436 ignition[787]: reading system config file "/usr/lib/ignition/user.ign" Mar 7 01:12:41.457446 ignition[787]: no config at "/usr/lib/ignition/user.ign" Mar 7 01:12:41.457486 ignition[787]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Mar 7 01:12:41.457685 ignition[787]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Mar 7 01:12:41.482592 systemd-networkd[780]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Mar 7 01:12:41.520621 systemd-networkd[780]: eth0: DHCPv4 address 95.216.146.133/32, gateway 172.31.1.1 acquired from 172.31.1.1 Mar 7 01:12:41.658178 ignition[787]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Mar 7 01:12:41.662343 ignition[787]: GET result: OK Mar 7 01:12:41.662593 ignition[787]: parsing config with SHA512: 1f2475365ec42e32e9db85d1083a2fc14db642f8c1e00fa1aa5bf2ec97918d42e2bf2db4fe6137d6043cc9632b6f4b47403f12fb41112bfbec0d5e8cca304da9 Mar 7 01:12:41.670445 unknown[787]: fetched base config from "system" Mar 7 01:12:41.671531 unknown[787]: fetched base config from "system" Mar 7 01:12:41.671552 unknown[787]: fetched user config from "hetzner" Mar 7 01:12:41.672183 ignition[787]: fetch: fetch complete Mar 7 01:12:41.672194 ignition[787]: fetch: fetch passed Mar 7 01:12:41.677283 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 7 01:12:41.672274 ignition[787]: Ignition finished successfully Mar 7 01:12:41.683703 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 7 01:12:41.721241 ignition[795]: Ignition 2.19.0 Mar 7 01:12:41.721262 ignition[795]: Stage: kargs Mar 7 01:12:41.721551 ignition[795]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:12:41.721573 ignition[795]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 01:12:41.725764 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 7 01:12:41.723175 ignition[795]: kargs: kargs passed Mar 7 01:12:41.723260 ignition[795]: Ignition finished successfully Mar 7 01:12:41.734742 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 7 01:12:41.755374 ignition[801]: Ignition 2.19.0 Mar 7 01:12:41.755988 ignition[801]: Stage: disks Mar 7 01:12:41.756133 ignition[801]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:12:41.756143 ignition[801]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 01:12:41.756697 ignition[801]: disks: disks passed Mar 7 01:12:41.756733 ignition[801]: Ignition finished successfully Mar 7 01:12:41.759194 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 7 01:12:41.760018 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 7 01:12:41.760829 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 7 01:12:41.761995 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 7 01:12:41.763149 systemd[1]: Reached target sysinit.target - System Initialization. Mar 7 01:12:41.764334 systemd[1]: Reached target basic.target - Basic System. Mar 7 01:12:41.770593 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 7 01:12:41.790853 systemd-fsck[809]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Mar 7 01:12:41.793910 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 7 01:12:41.798637 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 7 01:12:41.869486 kernel: EXT4-fs (sda9): mounted filesystem aab0506b-de72-4dd2-9393-24d7958f49a5 r/w with ordered data mode. Quota mode: none. Mar 7 01:12:41.869902 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 7 01:12:41.870759 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 7 01:12:41.876530 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 7 01:12:41.879531 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 7 01:12:41.880902 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Mar 7 01:12:41.881597 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 7 01:12:41.882244 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 7 01:12:41.889231 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 7 01:12:41.890690 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by mount (817) Mar 7 01:12:41.890712 kernel: BTRFS info (device sda6): first mount of filesystem 872bf425-12c9-4ef2-aaf0-71379b3513d9 Mar 7 01:12:41.890721 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 7 01:12:41.890729 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:12:41.904164 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 7 01:12:41.904195 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:12:41.908567 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 7 01:12:41.910529 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 7 01:12:41.938285 coreos-metadata[819]: Mar 07 01:12:41.938 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Mar 7 01:12:41.939748 coreos-metadata[819]: Mar 07 01:12:41.939 INFO Fetch successful Mar 7 01:12:41.940976 coreos-metadata[819]: Mar 07 01:12:41.940 INFO wrote hostname ci-4081-3-6-n-0c8881e772 to /sysroot/etc/hostname Mar 7 01:12:41.942928 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 7 01:12:41.947560 initrd-setup-root[846]: cut: /sysroot/etc/passwd: No such file or directory Mar 7 01:12:41.953561 initrd-setup-root[853]: cut: /sysroot/etc/group: No such file or directory Mar 7 01:12:41.958248 initrd-setup-root[860]: cut: /sysroot/etc/shadow: No such file or directory Mar 7 01:12:41.962299 initrd-setup-root[867]: cut: /sysroot/etc/gshadow: No such file or directory Mar 7 01:12:42.044924 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 7 01:12:42.048546 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 7 01:12:42.051575 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 7 01:12:42.063492 kernel: BTRFS info (device sda6): last unmount of filesystem 872bf425-12c9-4ef2-aaf0-71379b3513d9 Mar 7 01:12:42.077030 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 7 01:12:42.086239 ignition[935]: INFO : Ignition 2.19.0 Mar 7 01:12:42.086239 ignition[935]: INFO : Stage: mount Mar 7 01:12:42.087159 ignition[935]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:12:42.087159 ignition[935]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 01:12:42.088883 ignition[935]: INFO : mount: mount passed Mar 7 01:12:42.088883 ignition[935]: INFO : Ignition finished successfully Mar 7 01:12:42.088531 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 7 01:12:42.093545 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 7 01:12:42.260143 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 7 01:12:42.265864 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 7 01:12:42.276733 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (947) Mar 7 01:12:42.276782 kernel: BTRFS info (device sda6): first mount of filesystem 872bf425-12c9-4ef2-aaf0-71379b3513d9 Mar 7 01:12:42.281424 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 7 01:12:42.281488 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:12:42.292532 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 7 01:12:42.292594 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:12:42.297529 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 7 01:12:42.334683 ignition[964]: INFO : Ignition 2.19.0 Mar 7 01:12:42.334683 ignition[964]: INFO : Stage: files Mar 7 01:12:42.336409 ignition[964]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:12:42.336409 ignition[964]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 01:12:42.336409 ignition[964]: DEBUG : files: compiled without relabeling support, skipping Mar 7 01:12:42.336409 ignition[964]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 7 01:12:42.336409 ignition[964]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 7 01:12:42.339593 ignition[964]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 7 01:12:42.339593 ignition[964]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 7 01:12:42.339593 ignition[964]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 7 01:12:42.339571 unknown[964]: wrote ssh authorized keys file for user: core Mar 7 01:12:42.342424 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 7 01:12:42.342424 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 7 01:12:42.342424 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 7 01:12:42.342424 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Mar 7 01:12:42.564698 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 7 01:12:42.791953 systemd-networkd[780]: eth1: Gained IPv6LL Mar 7 01:12:42.870986 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 7 01:12:42.870986 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 7 01:12:42.874298 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-x86-64.raw: attempt #1 Mar 7 01:12:42.919857 systemd-networkd[780]: eth0: Gained IPv6LL Mar 7 01:12:43.289639 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Mar 7 01:12:43.566663 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 7 01:12:43.566663 ignition[964]: INFO : files: op(c): [started] processing unit "containerd.service" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(c): [finished] processing unit "containerd.service" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(10): [started] processing unit "coreos-metadata.service" Mar 7 01:12:43.567946 ignition[964]: INFO : files: op(10): op(11): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Mar 7 01:12:43.573363 ignition[964]: INFO : files: op(10): op(11): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Mar 7 01:12:43.573363 ignition[964]: INFO : files: op(10): [finished] processing unit "coreos-metadata.service" Mar 7 01:12:43.573363 ignition[964]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Mar 7 01:12:43.573363 ignition[964]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Mar 7 01:12:43.573363 ignition[964]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 7 01:12:43.573363 ignition[964]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 7 01:12:43.573363 ignition[964]: INFO : files: files passed Mar 7 01:12:43.573363 ignition[964]: INFO : Ignition finished successfully Mar 7 01:12:43.570362 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 7 01:12:43.575672 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 7 01:12:43.579149 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 7 01:12:43.579835 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 7 01:12:43.579922 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 7 01:12:43.593270 initrd-setup-root-after-ignition[992]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:12:43.593270 initrd-setup-root-after-ignition[992]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:12:43.594218 initrd-setup-root-after-ignition[996]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:12:43.595543 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 7 01:12:43.596178 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 7 01:12:43.599586 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 7 01:12:43.625362 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 7 01:12:43.625473 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 7 01:12:43.626823 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 7 01:12:43.627725 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 7 01:12:43.628198 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 7 01:12:43.629683 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 7 01:12:43.643624 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 7 01:12:43.648605 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 7 01:12:43.656155 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 7 01:12:43.656619 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 01:12:43.657123 systemd[1]: Stopped target timers.target - Timer Units. Mar 7 01:12:43.657879 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 7 01:12:43.657996 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 7 01:12:43.659004 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 7 01:12:43.659764 systemd[1]: Stopped target basic.target - Basic System. Mar 7 01:12:43.660404 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 7 01:12:43.661110 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 7 01:12:43.661819 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 7 01:12:43.662532 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 7 01:12:43.663245 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 7 01:12:43.664447 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 7 01:12:43.665144 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 7 01:12:43.665865 systemd[1]: Stopped target swap.target - Swaps. Mar 7 01:12:43.666555 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 7 01:12:43.666650 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 7 01:12:43.667651 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 7 01:12:43.668346 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 01:12:43.669043 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 7 01:12:43.669127 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 01:12:43.669758 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 7 01:12:43.669832 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 7 01:12:43.670840 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 7 01:12:43.670922 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 7 01:12:43.671608 systemd[1]: ignition-files.service: Deactivated successfully. Mar 7 01:12:43.671685 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 7 01:12:43.672290 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 7 01:12:43.672358 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 7 01:12:43.680663 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 7 01:12:43.681516 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 7 01:12:43.681618 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 01:12:43.685603 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 7 01:12:43.685975 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 7 01:12:43.686053 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 01:12:43.686893 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 7 01:12:43.686990 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 7 01:12:43.690932 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 7 01:12:43.691381 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 7 01:12:43.692356 ignition[1016]: INFO : Ignition 2.19.0 Mar 7 01:12:43.692356 ignition[1016]: INFO : Stage: umount Mar 7 01:12:43.693834 ignition[1016]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:12:43.693834 ignition[1016]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 01:12:43.696074 ignition[1016]: INFO : umount: umount passed Mar 7 01:12:43.696074 ignition[1016]: INFO : Ignition finished successfully Mar 7 01:12:43.696134 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 7 01:12:43.696248 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 7 01:12:43.697338 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 7 01:12:43.697408 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 7 01:12:43.698003 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 7 01:12:43.698041 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 7 01:12:43.698388 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 7 01:12:43.698422 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 7 01:12:43.698828 systemd[1]: Stopped target network.target - Network. Mar 7 01:12:43.699427 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 7 01:12:43.699489 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 7 01:12:43.700105 systemd[1]: Stopped target paths.target - Path Units. Mar 7 01:12:43.700899 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 7 01:12:43.704489 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 01:12:43.705868 systemd[1]: Stopped target slices.target - Slice Units. Mar 7 01:12:43.706548 systemd[1]: Stopped target sockets.target - Socket Units. Mar 7 01:12:43.707183 systemd[1]: iscsid.socket: Deactivated successfully. Mar 7 01:12:43.707223 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 7 01:12:43.707801 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 7 01:12:43.707844 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 7 01:12:43.708411 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 7 01:12:43.708448 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 7 01:12:43.709854 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 7 01:12:43.709891 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 7 01:12:43.710616 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 7 01:12:43.718811 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 7 01:12:43.720440 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 7 01:12:43.721493 systemd-networkd[780]: eth1: DHCPv6 lease lost Mar 7 01:12:43.727496 systemd-networkd[780]: eth0: DHCPv6 lease lost Mar 7 01:12:43.728773 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 7 01:12:43.728887 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 7 01:12:43.729785 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 7 01:12:43.729843 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 7 01:12:43.738589 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 7 01:12:43.738923 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 7 01:12:43.738968 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 7 01:12:43.739507 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 01:12:43.740552 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 7 01:12:43.740663 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 7 01:12:43.746770 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 7 01:12:43.746876 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 7 01:12:43.755153 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 7 01:12:43.755638 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 7 01:12:43.756027 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 7 01:12:43.756064 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 7 01:12:43.756693 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 7 01:12:43.756729 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 7 01:12:43.757311 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 7 01:12:43.757345 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 01:12:43.758212 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 7 01:12:43.758369 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 01:12:43.759269 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 7 01:12:43.759361 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 7 01:12:43.760696 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 7 01:12:43.760761 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 7 01:12:43.761321 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 7 01:12:43.761355 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 01:12:43.761957 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 7 01:12:43.761996 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 7 01:12:43.763076 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 7 01:12:43.763113 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 7 01:12:43.764099 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 01:12:43.764136 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:12:43.769627 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 7 01:12:43.769974 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 7 01:12:43.770027 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 01:12:43.770399 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:12:43.770432 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:12:43.775516 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 7 01:12:43.775613 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 7 01:12:43.776486 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 7 01:12:43.782578 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 7 01:12:43.790530 systemd[1]: Switching root. Mar 7 01:12:43.837842 systemd-journald[189]: Journal stopped Mar 7 01:12:44.937545 systemd-journald[189]: Received SIGTERM from PID 1 (systemd). Mar 7 01:12:44.937624 kernel: SELinux: policy capability network_peer_controls=1 Mar 7 01:12:44.937640 kernel: SELinux: policy capability open_perms=1 Mar 7 01:12:44.937653 kernel: SELinux: policy capability extended_socket_class=1 Mar 7 01:12:44.937672 kernel: SELinux: policy capability always_check_network=0 Mar 7 01:12:44.937684 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 7 01:12:44.937700 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 7 01:12:44.937711 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 7 01:12:44.937732 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 7 01:12:44.937742 kernel: audit: type=1403 audit(1772845964.010:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 7 01:12:44.937756 systemd[1]: Successfully loaded SELinux policy in 44.701ms. Mar 7 01:12:44.937774 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 18.877ms. Mar 7 01:12:44.937788 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 7 01:12:44.937801 systemd[1]: Detected virtualization kvm. Mar 7 01:12:44.937816 systemd[1]: Detected architecture x86-64. Mar 7 01:12:44.937825 systemd[1]: Detected first boot. Mar 7 01:12:44.937834 systemd[1]: Hostname set to . Mar 7 01:12:44.937842 systemd[1]: Initializing machine ID from VM UUID. Mar 7 01:12:44.937851 zram_generator::config[1075]: No configuration found. Mar 7 01:12:44.937861 systemd[1]: Populated /etc with preset unit settings. Mar 7 01:12:44.937874 systemd[1]: Queued start job for default target multi-user.target. Mar 7 01:12:44.937887 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Mar 7 01:12:44.937899 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 7 01:12:44.937910 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 7 01:12:44.937918 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 7 01:12:44.937927 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 7 01:12:44.937936 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 7 01:12:44.937948 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 7 01:12:44.937964 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 7 01:12:44.937973 systemd[1]: Created slice user.slice - User and Session Slice. Mar 7 01:12:44.937984 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 01:12:44.937993 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 01:12:44.938002 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 7 01:12:44.938011 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 7 01:12:44.938023 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 7 01:12:44.938037 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 7 01:12:44.938047 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 7 01:12:44.938056 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 01:12:44.938065 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 7 01:12:44.938076 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 01:12:44.938085 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 7 01:12:44.938095 systemd[1]: Reached target slices.target - Slice Units. Mar 7 01:12:44.938108 systemd[1]: Reached target swap.target - Swaps. Mar 7 01:12:44.938120 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 7 01:12:44.938131 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 7 01:12:44.938139 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 7 01:12:44.938151 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 7 01:12:44.938160 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 7 01:12:44.938169 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 7 01:12:44.938183 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 01:12:44.938196 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 7 01:12:44.938212 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 7 01:12:44.938224 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 7 01:12:44.938237 systemd[1]: Mounting media.mount - External Media Directory... Mar 7 01:12:44.938247 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:44.938260 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 7 01:12:44.938273 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 7 01:12:44.938286 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 7 01:12:44.938295 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 7 01:12:44.938304 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:12:44.938312 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 7 01:12:44.938321 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 7 01:12:44.938330 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:12:44.938343 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 7 01:12:44.938357 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:12:44.938370 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 7 01:12:44.938378 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:12:44.938387 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 7 01:12:44.938396 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Mar 7 01:12:44.938405 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Mar 7 01:12:44.938414 kernel: loop: module loaded Mar 7 01:12:44.938430 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 7 01:12:44.938443 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 7 01:12:44.938453 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 7 01:12:44.938590 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 7 01:12:44.938633 systemd-journald[1177]: Collecting audit messages is disabled. Mar 7 01:12:44.938666 kernel: ACPI: bus type drm_connector registered Mar 7 01:12:44.938682 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 7 01:12:44.938696 kernel: fuse: init (API version 7.39) Mar 7 01:12:44.938709 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:44.938725 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 7 01:12:44.938740 systemd-journald[1177]: Journal started Mar 7 01:12:44.938757 systemd-journald[1177]: Runtime Journal (/run/log/journal/39ce0ed0a9ee49bfabcd2a5641db3a8b) is 8.0M, max 76.3M, 68.3M free. Mar 7 01:12:44.941472 systemd[1]: Started systemd-journald.service - Journal Service. Mar 7 01:12:44.942239 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 7 01:12:44.942784 systemd[1]: Mounted media.mount - External Media Directory. Mar 7 01:12:44.943287 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 7 01:12:44.944132 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 7 01:12:44.944831 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 7 01:12:44.945607 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 7 01:12:44.946310 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 01:12:44.947035 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 7 01:12:44.947198 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 7 01:12:44.947996 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:12:44.948150 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:12:44.949011 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 7 01:12:44.949213 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 7 01:12:44.949912 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:12:44.950110 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:12:44.950865 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 7 01:12:44.951061 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 7 01:12:44.951742 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:12:44.951951 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:12:44.952822 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 7 01:12:44.953573 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 7 01:12:44.954269 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 7 01:12:44.966173 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 7 01:12:44.978552 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 7 01:12:44.980527 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 7 01:12:44.982323 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 7 01:12:44.991600 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 7 01:12:44.994120 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 7 01:12:44.998835 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 01:12:44.999985 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 7 01:12:45.003558 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 01:12:45.012648 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 7 01:12:45.019813 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 7 01:12:45.022260 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 7 01:12:45.023909 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 7 01:12:45.036546 systemd-journald[1177]: Time spent on flushing to /var/log/journal/39ce0ed0a9ee49bfabcd2a5641db3a8b is 35.958ms for 1163 entries. Mar 7 01:12:45.036546 systemd-journald[1177]: System Journal (/var/log/journal/39ce0ed0a9ee49bfabcd2a5641db3a8b) is 8.0M, max 584.8M, 576.8M free. Mar 7 01:12:45.096542 systemd-journald[1177]: Received client request to flush runtime journal. Mar 7 01:12:45.045404 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 7 01:12:45.045925 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 7 01:12:45.068206 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 7 01:12:45.090723 systemd-tmpfiles[1219]: ACLs are not supported, ignoring. Mar 7 01:12:45.090734 systemd-tmpfiles[1219]: ACLs are not supported, ignoring. Mar 7 01:12:45.099976 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 7 01:12:45.106946 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 7 01:12:45.119641 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 7 01:12:45.143852 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 01:12:45.154617 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 7 01:12:45.164704 udevadm[1238]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Mar 7 01:12:45.171879 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 7 01:12:45.183591 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 7 01:12:45.197070 systemd-tmpfiles[1241]: ACLs are not supported, ignoring. Mar 7 01:12:45.197316 systemd-tmpfiles[1241]: ACLs are not supported, ignoring. Mar 7 01:12:45.201764 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 01:12:45.401418 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 7 01:12:45.410727 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 01:12:45.437787 systemd-udevd[1247]: Using default interface naming scheme 'v255'. Mar 7 01:12:45.474023 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 01:12:45.481715 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 7 01:12:45.496594 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 7 01:12:45.552185 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Mar 7 01:12:45.553882 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 7 01:12:45.583701 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 34 scanned by (udev-worker) (1252) Mar 7 01:12:45.630767 systemd-networkd[1251]: lo: Link UP Mar 7 01:12:45.631527 systemd-networkd[1251]: lo: Gained carrier Mar 7 01:12:45.636019 systemd-networkd[1251]: Enumeration completed Mar 7 01:12:45.636432 systemd-networkd[1251]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:45.636438 systemd-networkd[1251]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 01:12:45.637221 systemd-networkd[1251]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:45.637254 systemd-networkd[1251]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 01:12:45.637610 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 7 01:12:45.639101 systemd-networkd[1251]: eth0: Link UP Mar 7 01:12:45.639143 systemd-networkd[1251]: eth0: Gained carrier Mar 7 01:12:45.639176 systemd-networkd[1251]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:45.645765 systemd-networkd[1251]: eth1: Link UP Mar 7 01:12:45.645828 systemd-networkd[1251]: eth1: Gained carrier Mar 7 01:12:45.645844 systemd-networkd[1251]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:45.648576 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 7 01:12:45.663499 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input4 Mar 7 01:12:45.672622 systemd-networkd[1251]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Mar 7 01:12:45.681481 kernel: ACPI: button: Power Button [PWRF] Mar 7 01:12:45.688841 systemd-networkd[1251]: eth0: DHCPv4 address 95.216.146.133/32, gateway 172.31.1.1 acquired from 172.31.1.1 Mar 7 01:12:45.692356 systemd-networkd[1251]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:45.703200 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:45.703321 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:12:45.710619 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:12:45.712574 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:12:45.722195 systemd-networkd[1251]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:12:45.723622 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:12:45.724033 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 7 01:12:45.724064 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 7 01:12:45.724094 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:45.727723 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Mar 7 01:12:45.729674 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 7 01:12:45.736905 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:12:45.737096 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:12:45.738029 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 01:12:45.739476 kernel: mousedev: PS/2 mouse device common for all mice Mar 7 01:12:45.748992 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:12:45.754781 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Mar 7 01:12:45.754997 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 7 01:12:45.754168 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:12:45.773651 kernel: EDAC MC: Ver: 3.0.0 Mar 7 01:12:45.783503 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input5 Mar 7 01:12:45.802503 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:12:45.805208 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:12:45.805422 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:12:45.812829 kernel: [drm] pci: virtio-vga detected at 0000:00:01.0 Mar 7 01:12:45.814809 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Mar 7 01:12:45.817660 kernel: Console: switching to colour dummy device 80x25 Mar 7 01:12:45.817614 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 01:12:45.822237 kernel: virtio-pci 0000:00:01.0: vgaarb: deactivate vga console Mar 7 01:12:45.822449 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Mar 7 01:12:45.822485 kernel: [drm] features: -context_init Mar 7 01:12:45.825102 kernel: [drm] number of scanouts: 1 Mar 7 01:12:45.825141 kernel: [drm] number of cap sets: 0 Mar 7 01:12:45.828477 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Mar 7 01:12:45.835991 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Mar 7 01:12:45.836049 kernel: Console: switching to colour frame buffer device 160x50 Mar 7 01:12:45.843483 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Mar 7 01:12:45.847390 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:12:45.847779 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:12:45.856585 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:12:45.896829 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:12:45.976230 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 7 01:12:45.990623 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 7 01:12:45.998575 lvm[1321]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 7 01:12:46.030251 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 7 01:12:46.030546 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 7 01:12:46.035637 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 7 01:12:46.040843 lvm[1324]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 7 01:12:46.066600 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 7 01:12:46.067389 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 7 01:12:46.067527 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 7 01:12:46.067550 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 7 01:12:46.067640 systemd[1]: Reached target machines.target - Containers. Mar 7 01:12:46.069351 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Mar 7 01:12:46.079587 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 7 01:12:46.081792 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 7 01:12:46.082058 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 01:12:46.085679 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 7 01:12:46.088594 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Mar 7 01:12:46.093597 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 7 01:12:46.095429 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 7 01:12:46.107250 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 7 01:12:46.117578 kernel: loop0: detected capacity change from 0 to 8 Mar 7 01:12:46.129483 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 7 01:12:46.134155 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 7 01:12:46.136194 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Mar 7 01:12:46.148485 kernel: loop1: detected capacity change from 0 to 228704 Mar 7 01:12:46.175480 kernel: loop2: detected capacity change from 0 to 140768 Mar 7 01:12:46.207485 kernel: loop3: detected capacity change from 0 to 142488 Mar 7 01:12:46.243498 kernel: loop4: detected capacity change from 0 to 8 Mar 7 01:12:46.247491 kernel: loop5: detected capacity change from 0 to 228704 Mar 7 01:12:46.260500 kernel: loop6: detected capacity change from 0 to 140768 Mar 7 01:12:46.276483 kernel: loop7: detected capacity change from 0 to 142488 Mar 7 01:12:46.293041 (sd-merge)[1346]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Mar 7 01:12:46.293736 (sd-merge)[1346]: Merged extensions into '/usr'. Mar 7 01:12:46.298036 systemd[1]: Reloading requested from client PID 1332 ('systemd-sysext') (unit systemd-sysext.service)... Mar 7 01:12:46.298116 systemd[1]: Reloading... Mar 7 01:12:46.373492 zram_generator::config[1374]: No configuration found. Mar 7 01:12:46.394984 ldconfig[1328]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 7 01:12:46.481813 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:12:46.548839 systemd[1]: Reloading finished in 250 ms. Mar 7 01:12:46.564417 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 7 01:12:46.567450 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 7 01:12:46.575585 systemd[1]: Starting ensure-sysext.service... Mar 7 01:12:46.577598 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 7 01:12:46.593548 systemd[1]: Reloading requested from client PID 1424 ('systemctl') (unit ensure-sysext.service)... Mar 7 01:12:46.593566 systemd[1]: Reloading... Mar 7 01:12:46.610882 systemd-tmpfiles[1425]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 7 01:12:46.611218 systemd-tmpfiles[1425]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 7 01:12:46.612138 systemd-tmpfiles[1425]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 7 01:12:46.612383 systemd-tmpfiles[1425]: ACLs are not supported, ignoring. Mar 7 01:12:46.612470 systemd-tmpfiles[1425]: ACLs are not supported, ignoring. Mar 7 01:12:46.616347 systemd-tmpfiles[1425]: Detected autofs mount point /boot during canonicalization of boot. Mar 7 01:12:46.616359 systemd-tmpfiles[1425]: Skipping /boot Mar 7 01:12:46.637315 systemd-tmpfiles[1425]: Detected autofs mount point /boot during canonicalization of boot. Mar 7 01:12:46.637590 systemd-tmpfiles[1425]: Skipping /boot Mar 7 01:12:46.647810 zram_generator::config[1451]: No configuration found. Mar 7 01:12:46.757124 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:12:46.815988 systemd[1]: Reloading finished in 221 ms. Mar 7 01:12:46.834537 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 01:12:46.846660 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 7 01:12:46.854576 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 7 01:12:46.864660 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 7 01:12:46.871594 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 7 01:12:46.875777 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 7 01:12:46.886734 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:46.887056 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:12:46.898708 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:12:46.903947 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:12:46.906232 augenrules[1525]: No rules Mar 7 01:12:46.916682 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:12:46.917235 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 01:12:46.917314 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:46.920723 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 7 01:12:46.921548 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 7 01:12:46.922260 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:12:46.922429 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:12:46.926184 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:12:46.926359 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:12:46.931591 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:12:46.931980 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:12:46.944909 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:46.945079 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:12:46.947639 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:12:46.957109 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:12:46.962683 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:12:46.963726 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 01:12:46.972818 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 7 01:12:46.973253 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:46.975737 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 7 01:12:46.978192 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:12:46.978361 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:12:46.979952 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:12:46.980120 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:12:46.980831 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:12:46.980983 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:12:46.989819 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 01:12:46.990048 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 01:12:46.996354 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:46.997415 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:12:47.004660 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:12:47.012727 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 7 01:12:47.015633 systemd-networkd[1251]: eth0: Gained IPv6LL Mar 7 01:12:47.020712 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:12:47.028712 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:12:47.029230 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 01:12:47.029320 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 7 01:12:47.033440 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 7 01:12:47.039024 systemd-resolved[1515]: Positive Trust Anchors: Mar 7 01:12:47.039325 systemd-resolved[1515]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 7 01:12:47.039386 systemd-resolved[1515]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 7 01:12:47.041060 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 7 01:12:47.044166 systemd-resolved[1515]: Using system hostname 'ci-4081-3-6-n-0c8881e772'. Mar 7 01:12:47.045520 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 7 01:12:47.046410 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:12:47.049237 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:12:47.051866 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 7 01:12:47.052974 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 7 01:12:47.053171 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 7 01:12:47.055079 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:12:47.055274 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:12:47.057263 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:12:47.057448 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:12:47.062094 systemd[1]: Finished ensure-sysext.service. Mar 7 01:12:47.070857 systemd[1]: Reached target network.target - Network. Mar 7 01:12:47.073322 systemd[1]: Reached target network-online.target - Network is Online. Mar 7 01:12:47.073732 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 7 01:12:47.074084 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 01:12:47.074143 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 01:12:47.079602 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 7 01:12:47.080530 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 7 01:12:47.135296 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 7 01:12:47.136793 systemd[1]: Reached target sysinit.target - System Initialization. Mar 7 01:12:47.137331 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 7 01:12:47.139189 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 7 01:12:47.139716 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 7 01:12:47.140187 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 7 01:12:47.140215 systemd[1]: Reached target paths.target - Path Units. Mar 7 01:12:47.142380 systemd[1]: Reached target time-set.target - System Time Set. Mar 7 01:12:47.143115 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 7 01:12:47.143716 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 7 01:12:47.144158 systemd[1]: Reached target timers.target - Timer Units. Mar 7 01:12:47.147637 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 7 01:12:47.149951 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 7 01:12:47.152879 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 7 01:12:47.154717 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 7 01:12:47.155296 systemd[1]: Reached target sockets.target - Socket Units. Mar 7 01:12:47.155621 systemd[1]: Reached target basic.target - Basic System. Mar 7 01:12:47.156103 systemd[1]: System is tainted: cgroupsv1 Mar 7 01:12:47.156131 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 7 01:12:47.156152 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 7 01:12:47.159372 systemd[1]: Starting containerd.service - containerd container runtime... Mar 7 01:12:47.170632 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 7 01:12:47.173525 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 7 01:12:47.178169 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 7 01:12:47.593702 systemd-timesyncd[1578]: Contacted time server 80.153.195.191:123 (0.flatcar.pool.ntp.org). Mar 7 01:12:47.593745 systemd-timesyncd[1578]: Initial clock synchronization to Sat 2026-03-07 01:12:47.593608 UTC. Mar 7 01:12:47.594664 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 7 01:12:47.595068 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 7 01:12:47.595513 systemd-resolved[1515]: Clock change detected. Flushing caches. Mar 7 01:12:47.600185 jq[1588]: false Mar 7 01:12:47.600489 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:12:47.603909 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 7 01:12:47.614577 coreos-metadata[1584]: Mar 07 01:12:47.614 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Mar 7 01:12:47.621222 coreos-metadata[1584]: Mar 07 01:12:47.616 INFO Fetch successful Mar 7 01:12:47.621222 coreos-metadata[1584]: Mar 07 01:12:47.616 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Mar 7 01:12:47.621222 coreos-metadata[1584]: Mar 07 01:12:47.617 INFO Fetch successful Mar 7 01:12:47.621171 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 7 01:12:47.631157 dbus-daemon[1585]: [system] SELinux support is enabled Mar 7 01:12:47.632767 extend-filesystems[1589]: Found loop4 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found loop5 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found loop6 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found loop7 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found sda Mar 7 01:12:47.638737 extend-filesystems[1589]: Found sda1 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found sda2 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found sda3 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found usr Mar 7 01:12:47.638737 extend-filesystems[1589]: Found sda4 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found sda6 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found sda7 Mar 7 01:12:47.638737 extend-filesystems[1589]: Found sda9 Mar 7 01:12:47.638737 extend-filesystems[1589]: Checking size of /dev/sda9 Mar 7 01:12:47.636332 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 7 01:12:47.643691 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Mar 7 01:12:47.663428 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 7 01:12:47.669597 extend-filesystems[1589]: Resized partition /dev/sda9 Mar 7 01:12:47.676263 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 7 01:12:47.679866 extend-filesystems[1616]: resize2fs 1.47.1 (20-May-2024) Mar 7 01:12:47.684678 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 7 01:12:47.686989 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 7 01:12:47.699564 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 19393531 blocks Mar 7 01:12:47.695033 systemd[1]: Starting update-engine.service - Update Engine... Mar 7 01:12:47.702204 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 7 01:12:47.704791 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 7 01:12:47.724613 update_engine[1622]: I20260307 01:12:47.722602 1622 main.cc:92] Flatcar Update Engine starting Mar 7 01:12:47.725995 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 7 01:12:47.726705 update_engine[1622]: I20260307 01:12:47.725617 1622 update_check_scheduler.cc:74] Next update check in 2m20s Mar 7 01:12:47.726267 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 7 01:12:47.737634 jq[1627]: true Mar 7 01:12:47.747727 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 34 scanned by (udev-worker) (1257) Mar 7 01:12:47.739947 systemd[1]: motdgen.service: Deactivated successfully. Mar 7 01:12:47.740203 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 7 01:12:47.748170 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 7 01:12:47.766337 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 7 01:12:47.767238 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 7 01:12:47.793818 (ntainerd)[1638]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 7 01:12:47.809682 systemd-networkd[1251]: eth1: Gained IPv6LL Mar 7 01:12:47.820257 tar[1635]: linux-amd64/LICENSE Mar 7 01:12:47.823041 jq[1637]: true Mar 7 01:12:47.829314 tar[1635]: linux-amd64/helm Mar 7 01:12:47.861876 systemd[1]: Started update-engine.service - Update Engine. Mar 7 01:12:47.863494 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 7 01:12:47.863520 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 7 01:12:47.863964 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 7 01:12:47.863978 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 7 01:12:47.867034 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 7 01:12:47.875243 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 7 01:12:47.886658 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 7 01:12:47.904403 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 7 01:12:47.913404 systemd-logind[1620]: New seat seat0. Mar 7 01:12:47.921586 systemd-logind[1620]: Watching system buttons on /dev/input/event2 (Power Button) Mar 7 01:12:47.921606 systemd-logind[1620]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 7 01:12:47.921815 systemd[1]: Started systemd-logind.service - User Login Management. Mar 7 01:12:47.952803 sshd_keygen[1629]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 7 01:12:47.979267 locksmithd[1663]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 7 01:12:47.992418 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 7 01:12:48.004100 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 7 01:12:48.011574 bash[1679]: Updated "/home/core/.ssh/authorized_keys" Mar 7 01:12:48.016040 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 7 01:12:48.030242 systemd[1]: Starting sshkeys.service... Mar 7 01:12:48.032737 systemd[1]: issuegen.service: Deactivated successfully. Mar 7 01:12:48.033159 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 7 01:12:48.035305 kernel: EXT4-fs (sda9): resized filesystem to 19393531 Mar 7 01:12:48.043785 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 7 01:12:48.051114 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Mar 7 01:12:48.058413 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Mar 7 01:12:48.061242 extend-filesystems[1616]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Mar 7 01:12:48.061242 extend-filesystems[1616]: old_desc_blocks = 1, new_desc_blocks = 10 Mar 7 01:12:48.061242 extend-filesystems[1616]: The filesystem on /dev/sda9 is now 19393531 (4k) blocks long. Mar 7 01:12:48.066016 extend-filesystems[1589]: Resized filesystem in /dev/sda9 Mar 7 01:12:48.066016 extend-filesystems[1589]: Found sr0 Mar 7 01:12:48.065880 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 7 01:12:48.066161 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 7 01:12:48.082447 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 7 01:12:48.094300 containerd[1638]: time="2026-03-07T01:12:48.094241963Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Mar 7 01:12:48.098888 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 7 01:12:48.111303 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 7 01:12:48.116124 systemd[1]: Reached target getty.target - Login Prompts. Mar 7 01:12:48.121324 coreos-metadata[1707]: Mar 07 01:12:48.120 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Mar 7 01:12:48.124505 coreos-metadata[1707]: Mar 07 01:12:48.124 INFO Fetch successful Mar 7 01:12:48.128393 unknown[1707]: wrote ssh authorized keys file for user: core Mar 7 01:12:48.132779 containerd[1638]: time="2026-03-07T01:12:48.131016698Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 7 01:12:48.151100 containerd[1638]: time="2026-03-07T01:12:48.150931427Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:12:48.151100 containerd[1638]: time="2026-03-07T01:12:48.150985827Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 7 01:12:48.151100 containerd[1638]: time="2026-03-07T01:12:48.151007067Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 7 01:12:48.151227 containerd[1638]: time="2026-03-07T01:12:48.151183977Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 7 01:12:48.151227 containerd[1638]: time="2026-03-07T01:12:48.151198547Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 7 01:12:48.151289 containerd[1638]: time="2026-03-07T01:12:48.151266447Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:12:48.151289 containerd[1638]: time="2026-03-07T01:12:48.151285297Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.151522847Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.151561727Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.151574387Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.151584297Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.151662517Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.151896867Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.152034667Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.152048107Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.152123687Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 7 01:12:48.153067 containerd[1638]: time="2026-03-07T01:12:48.152167237Z" level=info msg="metadata content store policy set" policy=shared Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158293500Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158340590Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158356510Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158371430Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158387160Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158523140Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158862610Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158972230Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158986130Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.158999210Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.159012380Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.159025940Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.159048550Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 7 01:12:48.159489 containerd[1638]: time="2026-03-07T01:12:48.159066510Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159081930Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159094580Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159106650Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159133110Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159153000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159168340Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159184950Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159203070Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159215680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159228880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159240270Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159257310Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159269680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.159829 containerd[1638]: time="2026-03-07T01:12:48.159283980Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159297240Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159309820Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159322600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159337860Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159356290Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159374990Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159388250Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159429570Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159445110Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159456630Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159468160Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159477590Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159489710Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 7 01:12:48.160055 containerd[1638]: time="2026-03-07T01:12:48.159499900Z" level=info msg="NRI interface is disabled by configuration." Mar 7 01:12:48.160277 containerd[1638]: time="2026-03-07T01:12:48.159511440Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.159774320Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.159835990Z" level=info msg="Connect containerd service" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.159869950Z" level=info msg="using legacy CRI server" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.159876970Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.159954880Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.160446371Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.160775741Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.160823301Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.161164611Z" level=info msg="Start subscribing containerd event" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.161224881Z" level=info msg="Start recovering state" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.161293791Z" level=info msg="Start event monitor" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.161305141Z" level=info msg="Start snapshots syncer" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.161313981Z" level=info msg="Start cni network conf syncer for default" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.161322201Z" level=info msg="Start streaming server" Mar 7 01:12:48.161710 containerd[1638]: time="2026-03-07T01:12:48.161642081Z" level=info msg="containerd successfully booted in 0.069263s" Mar 7 01:12:48.161478 systemd[1]: Started containerd.service - containerd container runtime. Mar 7 01:12:48.169522 update-ssh-keys[1721]: Updated "/home/core/.ssh/authorized_keys" Mar 7 01:12:48.177775 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Mar 7 01:12:48.182228 systemd[1]: Finished sshkeys.service. Mar 7 01:12:48.481902 tar[1635]: linux-amd64/README.md Mar 7 01:12:48.497572 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 7 01:12:48.700827 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:12:48.700994 (kubelet)[1742]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 01:12:48.702033 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 7 01:12:48.703182 systemd[1]: Startup finished in 6.741s (kernel) + 4.326s (userspace) = 11.067s. Mar 7 01:12:49.119865 kubelet[1742]: E0307 01:12:49.119806 1742 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 01:12:49.122729 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 01:12:49.122995 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 01:12:57.070285 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 7 01:12:57.076064 systemd[1]: Started sshd@0-95.216.146.133:22-4.153.228.146:49484.service - OpenSSH per-connection server daemon (4.153.228.146:49484). Mar 7 01:12:57.835577 sshd[1754]: Accepted publickey for core from 4.153.228.146 port 49484 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:12:57.838304 sshd[1754]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:12:57.854081 systemd-logind[1620]: New session 1 of user core. Mar 7 01:12:57.856727 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 7 01:12:57.862205 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 7 01:12:57.898719 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 7 01:12:57.910081 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 7 01:12:57.926969 (systemd)[1760]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 7 01:12:58.039830 systemd[1760]: Queued start job for default target default.target. Mar 7 01:12:58.040374 systemd[1760]: Created slice app.slice - User Application Slice. Mar 7 01:12:58.040392 systemd[1760]: Reached target paths.target - Paths. Mar 7 01:12:58.040403 systemd[1760]: Reached target timers.target - Timers. Mar 7 01:12:58.051622 systemd[1760]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 7 01:12:58.062900 systemd[1760]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 7 01:12:58.063031 systemd[1760]: Reached target sockets.target - Sockets. Mar 7 01:12:58.063080 systemd[1760]: Reached target basic.target - Basic System. Mar 7 01:12:58.063116 systemd[1760]: Reached target default.target - Main User Target. Mar 7 01:12:58.063145 systemd[1760]: Startup finished in 124ms. Mar 7 01:12:58.063371 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 7 01:12:58.069762 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 7 01:12:58.597233 systemd[1]: Started sshd@1-95.216.146.133:22-4.153.228.146:49492.service - OpenSSH per-connection server daemon (4.153.228.146:49492). Mar 7 01:12:59.137342 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 7 01:12:59.148913 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:12:59.284669 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:12:59.289394 (kubelet)[1785]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 01:12:59.325904 kubelet[1785]: E0307 01:12:59.325858 1785 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 01:12:59.330317 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 01:12:59.330527 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 01:12:59.363582 sshd[1772]: Accepted publickey for core from 4.153.228.146 port 49492 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:12:59.364273 sshd[1772]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:12:59.370195 systemd-logind[1620]: New session 2 of user core. Mar 7 01:12:59.381845 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 7 01:12:59.886582 sshd[1772]: pam_unix(sshd:session): session closed for user core Mar 7 01:12:59.891932 systemd-logind[1620]: Session 2 logged out. Waiting for processes to exit. Mar 7 01:12:59.893412 systemd[1]: sshd@1-95.216.146.133:22-4.153.228.146:49492.service: Deactivated successfully. Mar 7 01:12:59.897087 systemd[1]: session-2.scope: Deactivated successfully. Mar 7 01:12:59.898384 systemd-logind[1620]: Removed session 2. Mar 7 01:13:00.013918 systemd[1]: Started sshd@2-95.216.146.133:22-4.153.228.146:51224.service - OpenSSH per-connection server daemon (4.153.228.146:51224). Mar 7 01:13:00.756726 sshd[1800]: Accepted publickey for core from 4.153.228.146 port 51224 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:13:00.757960 sshd[1800]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:13:00.764016 systemd-logind[1620]: New session 3 of user core. Mar 7 01:13:00.781022 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 7 01:13:01.274038 sshd[1800]: pam_unix(sshd:session): session closed for user core Mar 7 01:13:01.278126 systemd[1]: sshd@2-95.216.146.133:22-4.153.228.146:51224.service: Deactivated successfully. Mar 7 01:13:01.283340 systemd-logind[1620]: Session 3 logged out. Waiting for processes to exit. Mar 7 01:13:01.284261 systemd[1]: session-3.scope: Deactivated successfully. Mar 7 01:13:01.287129 systemd-logind[1620]: Removed session 3. Mar 7 01:13:01.408992 systemd[1]: Started sshd@3-95.216.146.133:22-4.153.228.146:51234.service - OpenSSH per-connection server daemon (4.153.228.146:51234). Mar 7 01:13:02.149573 sshd[1808]: Accepted publickey for core from 4.153.228.146 port 51234 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:13:02.151772 sshd[1808]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:13:02.160994 systemd-logind[1620]: New session 4 of user core. Mar 7 01:13:02.175134 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 7 01:13:02.680079 sshd[1808]: pam_unix(sshd:session): session closed for user core Mar 7 01:13:02.685492 systemd[1]: sshd@3-95.216.146.133:22-4.153.228.146:51234.service: Deactivated successfully. Mar 7 01:13:02.691625 systemd-logind[1620]: Session 4 logged out. Waiting for processes to exit. Mar 7 01:13:02.693465 systemd[1]: session-4.scope: Deactivated successfully. Mar 7 01:13:02.697665 systemd-logind[1620]: Removed session 4. Mar 7 01:13:02.807275 systemd[1]: Started sshd@4-95.216.146.133:22-4.153.228.146:51244.service - OpenSSH per-connection server daemon (4.153.228.146:51244). Mar 7 01:13:03.575894 sshd[1816]: Accepted publickey for core from 4.153.228.146 port 51244 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:13:03.577697 sshd[1816]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:13:03.585356 systemd-logind[1620]: New session 5 of user core. Mar 7 01:13:03.597098 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 7 01:13:03.999273 sudo[1820]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 7 01:13:04.000041 sudo[1820]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 01:13:04.025922 sudo[1820]: pam_unix(sudo:session): session closed for user root Mar 7 01:13:04.147809 sshd[1816]: pam_unix(sshd:session): session closed for user core Mar 7 01:13:04.151697 systemd[1]: sshd@4-95.216.146.133:22-4.153.228.146:51244.service: Deactivated successfully. Mar 7 01:13:04.155120 systemd-logind[1620]: Session 5 logged out. Waiting for processes to exit. Mar 7 01:13:04.155282 systemd[1]: session-5.scope: Deactivated successfully. Mar 7 01:13:04.157901 systemd-logind[1620]: Removed session 5. Mar 7 01:13:04.276932 systemd[1]: Started sshd@5-95.216.146.133:22-4.153.228.146:51254.service - OpenSSH per-connection server daemon (4.153.228.146:51254). Mar 7 01:13:05.023292 sshd[1825]: Accepted publickey for core from 4.153.228.146 port 51254 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:13:05.026081 sshd[1825]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:13:05.034622 systemd-logind[1620]: New session 6 of user core. Mar 7 01:13:05.041089 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 7 01:13:05.432858 sudo[1830]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 7 01:13:05.433666 sudo[1830]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 01:13:05.439942 sudo[1830]: pam_unix(sudo:session): session closed for user root Mar 7 01:13:05.452034 sudo[1829]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Mar 7 01:13:05.452743 sudo[1829]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 01:13:05.472913 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Mar 7 01:13:05.490205 auditctl[1833]: No rules Mar 7 01:13:05.491086 systemd[1]: audit-rules.service: Deactivated successfully. Mar 7 01:13:05.491669 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Mar 7 01:13:05.507170 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 7 01:13:05.546870 augenrules[1852]: No rules Mar 7 01:13:05.549007 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 7 01:13:05.551868 sudo[1829]: pam_unix(sudo:session): session closed for user root Mar 7 01:13:05.670688 sshd[1825]: pam_unix(sshd:session): session closed for user core Mar 7 01:13:05.679217 systemd[1]: sshd@5-95.216.146.133:22-4.153.228.146:51254.service: Deactivated successfully. Mar 7 01:13:05.684309 systemd-logind[1620]: Session 6 logged out. Waiting for processes to exit. Mar 7 01:13:05.685459 systemd[1]: session-6.scope: Deactivated successfully. Mar 7 01:13:05.687406 systemd-logind[1620]: Removed session 6. Mar 7 01:13:05.808384 systemd[1]: Started sshd@6-95.216.146.133:22-4.153.228.146:51270.service - OpenSSH per-connection server daemon (4.153.228.146:51270). Mar 7 01:13:06.557759 sshd[1861]: Accepted publickey for core from 4.153.228.146 port 51270 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:13:06.560593 sshd[1861]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:13:06.569278 systemd-logind[1620]: New session 7 of user core. Mar 7 01:13:06.579132 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 7 01:13:06.968514 sudo[1865]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 7 01:13:06.969446 sudo[1865]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 01:13:07.328888 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 7 01:13:07.329862 (dockerd)[1881]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 7 01:13:07.551585 dockerd[1881]: time="2026-03-07T01:13:07.551496997Z" level=info msg="Starting up" Mar 7 01:13:07.691867 systemd[1]: var-lib-docker-metacopy\x2dcheck977486376-merged.mount: Deactivated successfully. Mar 7 01:13:07.709826 dockerd[1881]: time="2026-03-07T01:13:07.709782153Z" level=info msg="Loading containers: start." Mar 7 01:13:07.807571 kernel: Initializing XFRM netlink socket Mar 7 01:13:07.884955 systemd-networkd[1251]: docker0: Link UP Mar 7 01:13:07.901969 dockerd[1881]: time="2026-03-07T01:13:07.901919483Z" level=info msg="Loading containers: done." Mar 7 01:13:07.920200 dockerd[1881]: time="2026-03-07T01:13:07.920148450Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 7 01:13:07.920325 dockerd[1881]: time="2026-03-07T01:13:07.920257860Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Mar 7 01:13:07.920433 dockerd[1881]: time="2026-03-07T01:13:07.920398190Z" level=info msg="Daemon has completed initialization" Mar 7 01:13:07.956170 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 7 01:13:07.956883 dockerd[1881]: time="2026-03-07T01:13:07.956592836Z" level=info msg="API listen on /run/docker.sock" Mar 7 01:13:08.343936 containerd[1638]: time="2026-03-07T01:13:08.343740457Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\"" Mar 7 01:13:08.640259 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4289948418-merged.mount: Deactivated successfully. Mar 7 01:13:09.008586 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2508172982.mount: Deactivated successfully. Mar 7 01:13:09.387869 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 7 01:13:09.397291 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:13:09.557668 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:13:09.568876 (kubelet)[2088]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 01:13:09.593679 kubelet[2088]: E0307 01:13:09.593624 2088 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 01:13:09.596341 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 01:13:09.596575 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 01:13:10.142372 containerd[1638]: time="2026-03-07T01:13:10.142329926Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:10.143206 containerd[1638]: time="2026-03-07T01:13:10.143137416Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.9: active requests=0, bytes read=30116286" Mar 7 01:13:10.144935 containerd[1638]: time="2026-03-07T01:13:10.143874346Z" level=info msg="ImageCreate event name:\"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:10.145947 containerd[1638]: time="2026-03-07T01:13:10.145607567Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:10.146915 containerd[1638]: time="2026-03-07T01:13:10.146376798Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.9\" with image id \"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\", size \"30112785\" in 1.802604471s" Mar 7 01:13:10.146915 containerd[1638]: time="2026-03-07T01:13:10.146400978Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\" returns image reference \"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\"" Mar 7 01:13:10.147207 containerd[1638]: time="2026-03-07T01:13:10.147187648Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\"" Mar 7 01:13:11.399176 containerd[1638]: time="2026-03-07T01:13:11.399133229Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:11.400196 containerd[1638]: time="2026-03-07T01:13:11.400013240Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.9: active requests=0, bytes read=26021832" Mar 7 01:13:11.401830 containerd[1638]: time="2026-03-07T01:13:11.401006460Z" level=info msg="ImageCreate event name:\"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:11.402940 containerd[1638]: time="2026-03-07T01:13:11.402923621Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:11.403625 containerd[1638]: time="2026-03-07T01:13:11.403608771Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.9\" with image id \"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\", size \"27678758\" in 1.256400843s" Mar 7 01:13:11.403691 containerd[1638]: time="2026-03-07T01:13:11.403680721Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\" returns image reference \"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\"" Mar 7 01:13:11.404292 containerd[1638]: time="2026-03-07T01:13:11.404272431Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\"" Mar 7 01:13:12.521861 containerd[1638]: time="2026-03-07T01:13:12.521800667Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:12.522786 containerd[1638]: time="2026-03-07T01:13:12.522661637Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.9: active requests=0, bytes read=20162768" Mar 7 01:13:12.523667 containerd[1638]: time="2026-03-07T01:13:12.523407218Z" level=info msg="ImageCreate event name:\"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:12.525228 containerd[1638]: time="2026-03-07T01:13:12.525202438Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:12.526076 containerd[1638]: time="2026-03-07T01:13:12.525958379Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.9\" with image id \"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\", size \"21819712\" in 1.121621658s" Mar 7 01:13:12.526076 containerd[1638]: time="2026-03-07T01:13:12.525980219Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\" returns image reference \"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\"" Mar 7 01:13:12.526369 containerd[1638]: time="2026-03-07T01:13:12.526274969Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\"" Mar 7 01:13:13.557642 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3074376658.mount: Deactivated successfully. Mar 7 01:13:13.909649 containerd[1638]: time="2026-03-07T01:13:13.909606635Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:13.910778 containerd[1638]: time="2026-03-07T01:13:13.910653855Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.9: active requests=0, bytes read=31828675" Mar 7 01:13:13.912332 containerd[1638]: time="2026-03-07T01:13:13.911514926Z" level=info msg="ImageCreate event name:\"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:13.913563 containerd[1638]: time="2026-03-07T01:13:13.913107666Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:13.913620 containerd[1638]: time="2026-03-07T01:13:13.913525556Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.9\" with image id \"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\", repo tag \"registry.k8s.io/kube-proxy:v1.33.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\", size \"31827666\" in 1.387193187s" Mar 7 01:13:13.913675 containerd[1638]: time="2026-03-07T01:13:13.913663227Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\" returns image reference \"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\"" Mar 7 01:13:13.914145 containerd[1638]: time="2026-03-07T01:13:13.914116147Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Mar 7 01:13:14.434502 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount253821685.mount: Deactivated successfully. Mar 7 01:13:15.224651 containerd[1638]: time="2026-03-07T01:13:15.224601363Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:15.225518 containerd[1638]: time="2026-03-07T01:13:15.225483733Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942332" Mar 7 01:13:15.226320 containerd[1638]: time="2026-03-07T01:13:15.226006803Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:15.228105 containerd[1638]: time="2026-03-07T01:13:15.228081684Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:15.228976 containerd[1638]: time="2026-03-07T01:13:15.228946624Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.314807637s" Mar 7 01:13:15.229036 containerd[1638]: time="2026-03-07T01:13:15.229024454Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Mar 7 01:13:15.229820 containerd[1638]: time="2026-03-07T01:13:15.229794425Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 7 01:13:15.708662 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount250203422.mount: Deactivated successfully. Mar 7 01:13:15.716591 containerd[1638]: time="2026-03-07T01:13:15.715116917Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:15.716591 containerd[1638]: time="2026-03-07T01:13:15.716268697Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321160" Mar 7 01:13:15.717282 containerd[1638]: time="2026-03-07T01:13:15.717236008Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:15.720861 containerd[1638]: time="2026-03-07T01:13:15.720810769Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:15.722318 containerd[1638]: time="2026-03-07T01:13:15.722265690Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 492.443175ms" Mar 7 01:13:15.722431 containerd[1638]: time="2026-03-07T01:13:15.722317440Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Mar 7 01:13:15.723841 containerd[1638]: time="2026-03-07T01:13:15.723798020Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Mar 7 01:13:16.235111 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount595262698.mount: Deactivated successfully. Mar 7 01:13:16.966668 containerd[1638]: time="2026-03-07T01:13:16.966624488Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:16.967410 containerd[1638]: time="2026-03-07T01:13:16.967382858Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=23718940" Mar 7 01:13:16.968557 containerd[1638]: time="2026-03-07T01:13:16.968500169Z" level=info msg="ImageCreate event name:\"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:16.970564 containerd[1638]: time="2026-03-07T01:13:16.970481000Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:16.971731 containerd[1638]: time="2026-03-07T01:13:16.971217920Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"23716032\" in 1.24738248s" Mar 7 01:13:16.971731 containerd[1638]: time="2026-03-07T01:13:16.971241980Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\"" Mar 7 01:13:18.934060 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:13:18.940162 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:13:18.974226 systemd[1]: Reloading requested from client PID 2264 ('systemctl') (unit session-7.scope)... Mar 7 01:13:18.974253 systemd[1]: Reloading... Mar 7 01:13:19.086654 zram_generator::config[2305]: No configuration found. Mar 7 01:13:19.191391 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:13:19.257270 systemd[1]: Reloading finished in 282 ms. Mar 7 01:13:19.314996 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 7 01:13:19.315109 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 7 01:13:19.315559 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:13:19.321835 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:13:19.470701 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:13:19.483238 (kubelet)[2369]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 7 01:13:19.514242 kubelet[2369]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 01:13:19.514242 kubelet[2369]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 7 01:13:19.514242 kubelet[2369]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 01:13:19.514593 kubelet[2369]: I0307 01:13:19.514264 2369 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 7 01:13:19.613019 kubelet[2369]: I0307 01:13:19.612971 2369 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 7 01:13:19.613019 kubelet[2369]: I0307 01:13:19.612993 2369 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 7 01:13:19.613212 kubelet[2369]: I0307 01:13:19.613139 2369 server.go:956] "Client rotation is on, will bootstrap in background" Mar 7 01:13:19.634080 kubelet[2369]: E0307 01:13:19.634040 2369 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://95.216.146.133:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 7 01:13:19.638097 kubelet[2369]: I0307 01:13:19.637654 2369 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 7 01:13:19.641446 kubelet[2369]: E0307 01:13:19.641416 2369 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 7 01:13:19.641446 kubelet[2369]: I0307 01:13:19.641442 2369 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 7 01:13:19.647771 kubelet[2369]: I0307 01:13:19.647755 2369 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 7 01:13:19.648724 kubelet[2369]: I0307 01:13:19.648681 2369 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 7 01:13:19.648831 kubelet[2369]: I0307 01:13:19.648710 2369 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-6-n-0c8881e772","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Mar 7 01:13:19.648899 kubelet[2369]: I0307 01:13:19.648838 2369 topology_manager.go:138] "Creating topology manager with none policy" Mar 7 01:13:19.648899 kubelet[2369]: I0307 01:13:19.648847 2369 container_manager_linux.go:303] "Creating device plugin manager" Mar 7 01:13:19.648984 kubelet[2369]: I0307 01:13:19.648966 2369 state_mem.go:36] "Initialized new in-memory state store" Mar 7 01:13:19.652424 kubelet[2369]: I0307 01:13:19.652400 2369 kubelet.go:480] "Attempting to sync node with API server" Mar 7 01:13:19.652424 kubelet[2369]: I0307 01:13:19.652417 2369 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 7 01:13:19.652483 kubelet[2369]: I0307 01:13:19.652440 2369 kubelet.go:386] "Adding apiserver pod source" Mar 7 01:13:19.652483 kubelet[2369]: I0307 01:13:19.652462 2369 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 7 01:13:19.657494 kubelet[2369]: E0307 01:13:19.656698 2369 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://95.216.146.133:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-6-n-0c8881e772&limit=500&resourceVersion=0\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 7 01:13:19.657494 kubelet[2369]: E0307 01:13:19.656750 2369 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://95.216.146.133:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 7 01:13:19.657494 kubelet[2369]: I0307 01:13:19.657015 2369 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 7 01:13:19.657494 kubelet[2369]: I0307 01:13:19.657412 2369 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 7 01:13:19.658554 kubelet[2369]: W0307 01:13:19.657962 2369 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 7 01:13:19.662048 kubelet[2369]: I0307 01:13:19.661847 2369 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 7 01:13:19.662048 kubelet[2369]: I0307 01:13:19.661879 2369 server.go:1289] "Started kubelet" Mar 7 01:13:19.662756 kubelet[2369]: I0307 01:13:19.662423 2369 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 7 01:13:19.663053 kubelet[2369]: I0307 01:13:19.663025 2369 server.go:317] "Adding debug handlers to kubelet server" Mar 7 01:13:19.664148 kubelet[2369]: I0307 01:13:19.663731 2369 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 7 01:13:19.664148 kubelet[2369]: I0307 01:13:19.663998 2369 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 7 01:13:19.666393 kubelet[2369]: E0307 01:13:19.664068 2369 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://95.216.146.133:6443/api/v1/namespaces/default/events\": dial tcp 95.216.146.133:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081-3-6-n-0c8881e772.189a6a0b335156cb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-3-6-n-0c8881e772,UID:ci-4081-3-6-n-0c8881e772,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-3-6-n-0c8881e772,},FirstTimestamp:2026-03-07 01:13:19.661864651 +0000 UTC m=+0.173010013,LastTimestamp:2026-03-07 01:13:19.661864651 +0000 UTC m=+0.173010013,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-6-n-0c8881e772,}" Mar 7 01:13:19.667420 kubelet[2369]: I0307 01:13:19.667009 2369 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 7 01:13:19.668162 kubelet[2369]: E0307 01:13:19.668150 2369 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 7 01:13:19.668621 kubelet[2369]: I0307 01:13:19.668609 2369 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 7 01:13:19.671563 kubelet[2369]: E0307 01:13:19.670820 2369 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-0c8881e772\" not found" Mar 7 01:13:19.671563 kubelet[2369]: I0307 01:13:19.670851 2369 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 7 01:13:19.671563 kubelet[2369]: I0307 01:13:19.670970 2369 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 7 01:13:19.671563 kubelet[2369]: I0307 01:13:19.671002 2369 reconciler.go:26] "Reconciler: start to sync state" Mar 7 01:13:19.671563 kubelet[2369]: E0307 01:13:19.671208 2369 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://95.216.146.133:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 7 01:13:19.671563 kubelet[2369]: E0307 01:13:19.671377 2369 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://95.216.146.133:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-0c8881e772?timeout=10s\": dial tcp 95.216.146.133:6443: connect: connection refused" interval="200ms" Mar 7 01:13:19.671563 kubelet[2369]: I0307 01:13:19.671472 2369 factory.go:223] Registration of the systemd container factory successfully Mar 7 01:13:19.671563 kubelet[2369]: I0307 01:13:19.671520 2369 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 7 01:13:19.675281 kubelet[2369]: I0307 01:13:19.675269 2369 factory.go:223] Registration of the containerd container factory successfully Mar 7 01:13:19.695131 kubelet[2369]: I0307 01:13:19.695088 2369 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 7 01:13:19.699809 kubelet[2369]: I0307 01:13:19.699787 2369 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 7 01:13:19.699809 kubelet[2369]: I0307 01:13:19.699806 2369 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 7 01:13:19.699873 kubelet[2369]: I0307 01:13:19.699826 2369 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 7 01:13:19.699873 kubelet[2369]: I0307 01:13:19.699834 2369 kubelet.go:2436] "Starting kubelet main sync loop" Mar 7 01:13:19.699914 kubelet[2369]: E0307 01:13:19.699869 2369 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 7 01:13:19.704405 kubelet[2369]: E0307 01:13:19.704385 2369 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://95.216.146.133:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 7 01:13:19.708059 kubelet[2369]: I0307 01:13:19.708048 2369 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 7 01:13:19.708147 kubelet[2369]: I0307 01:13:19.708139 2369 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 7 01:13:19.708185 kubelet[2369]: I0307 01:13:19.708180 2369 state_mem.go:36] "Initialized new in-memory state store" Mar 7 01:13:19.710864 kubelet[2369]: I0307 01:13:19.710853 2369 policy_none.go:49] "None policy: Start" Mar 7 01:13:19.710941 kubelet[2369]: I0307 01:13:19.710933 2369 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 7 01:13:19.710984 kubelet[2369]: I0307 01:13:19.710979 2369 state_mem.go:35] "Initializing new in-memory state store" Mar 7 01:13:19.715579 kubelet[2369]: E0307 01:13:19.715566 2369 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 7 01:13:19.715769 kubelet[2369]: I0307 01:13:19.715760 2369 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 7 01:13:19.715826 kubelet[2369]: I0307 01:13:19.715804 2369 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 7 01:13:19.717322 kubelet[2369]: I0307 01:13:19.717311 2369 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 7 01:13:19.719051 kubelet[2369]: E0307 01:13:19.719038 2369 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 7 01:13:19.719105 kubelet[2369]: E0307 01:13:19.719098 2369 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081-3-6-n-0c8881e772\" not found" Mar 7 01:13:19.814473 kubelet[2369]: E0307 01:13:19.814319 2369 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.821862 kubelet[2369]: E0307 01:13:19.821011 2369 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.826736 kubelet[2369]: E0307 01:13:19.826668 2369 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.827649 kubelet[2369]: I0307 01:13:19.827097 2369 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.828031 kubelet[2369]: E0307 01:13:19.827968 2369 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://95.216.146.133:6443/api/v1/nodes\": dial tcp 95.216.146.133:6443: connect: connection refused" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872416 kubelet[2369]: I0307 01:13:19.872355 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4a90aea135cdd8a005d68e1fd7967f4d-ca-certs\") pod \"kube-apiserver-ci-4081-3-6-n-0c8881e772\" (UID: \"4a90aea135cdd8a005d68e1fd7967f4d\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872416 kubelet[2369]: I0307 01:13:19.872420 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4a90aea135cdd8a005d68e1fd7967f4d-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-6-n-0c8881e772\" (UID: \"4a90aea135cdd8a005d68e1fd7967f4d\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872689 kubelet[2369]: I0307 01:13:19.872453 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-ca-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872689 kubelet[2369]: I0307 01:13:19.872477 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872689 kubelet[2369]: I0307 01:13:19.872522 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872689 kubelet[2369]: I0307 01:13:19.872579 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872689 kubelet[2369]: I0307 01:13:19.872615 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/cf82c8171bce44ee9aaf14eedd4a2d4c-kubeconfig\") pod \"kube-scheduler-ci-4081-3-6-n-0c8881e772\" (UID: \"cf82c8171bce44ee9aaf14eedd4a2d4c\") " pod="kube-system/kube-scheduler-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872993 kubelet[2369]: I0307 01:13:19.872651 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4a90aea135cdd8a005d68e1fd7967f4d-k8s-certs\") pod \"kube-apiserver-ci-4081-3-6-n-0c8881e772\" (UID: \"4a90aea135cdd8a005d68e1fd7967f4d\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.872993 kubelet[2369]: I0307 01:13:19.872680 2369 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:19.873602 kubelet[2369]: E0307 01:13:19.873530 2369 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://95.216.146.133:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-0c8881e772?timeout=10s\": dial tcp 95.216.146.133:6443: connect: connection refused" interval="400ms" Mar 7 01:13:20.031526 kubelet[2369]: I0307 01:13:20.031474 2369 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:20.031986 kubelet[2369]: E0307 01:13:20.031945 2369 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://95.216.146.133:6443/api/v1/nodes\": dial tcp 95.216.146.133:6443: connect: connection refused" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:20.119054 containerd[1638]: time="2026-03-07T01:13:20.118859721Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-6-n-0c8881e772,Uid:4a90aea135cdd8a005d68e1fd7967f4d,Namespace:kube-system,Attempt:0,}" Mar 7 01:13:20.123298 containerd[1638]: time="2026-03-07T01:13:20.122221721Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-6-n-0c8881e772,Uid:a160ce3374f2ce383338f0b710d12611,Namespace:kube-system,Attempt:0,}" Mar 7 01:13:20.129042 containerd[1638]: time="2026-03-07T01:13:20.128970161Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-6-n-0c8881e772,Uid:cf82c8171bce44ee9aaf14eedd4a2d4c,Namespace:kube-system,Attempt:0,}" Mar 7 01:13:20.275249 kubelet[2369]: E0307 01:13:20.275162 2369 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://95.216.146.133:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-0c8881e772?timeout=10s\": dial tcp 95.216.146.133:6443: connect: connection refused" interval="800ms" Mar 7 01:13:20.435936 kubelet[2369]: I0307 01:13:20.435854 2369 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:20.436658 kubelet[2369]: E0307 01:13:20.436492 2369 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://95.216.146.133:6443/api/v1/nodes\": dial tcp 95.216.146.133:6443: connect: connection refused" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:20.601367 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3391246129.mount: Deactivated successfully. Mar 7 01:13:20.607948 containerd[1638]: time="2026-03-07T01:13:20.607852244Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:13:20.609923 containerd[1638]: time="2026-03-07T01:13:20.609828771Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 7 01:13:20.610605 containerd[1638]: time="2026-03-07T01:13:20.610523138Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:13:20.611824 containerd[1638]: time="2026-03-07T01:13:20.611753920Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:13:20.613448 containerd[1638]: time="2026-03-07T01:13:20.613301111Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312078" Mar 7 01:13:20.614637 containerd[1638]: time="2026-03-07T01:13:20.614421684Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 7 01:13:20.614637 containerd[1638]: time="2026-03-07T01:13:20.614532713Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:13:20.619633 containerd[1638]: time="2026-03-07T01:13:20.619575614Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:13:20.622620 containerd[1638]: time="2026-03-07T01:13:20.622188429Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 503.196669ms" Mar 7 01:13:20.624692 containerd[1638]: time="2026-03-07T01:13:20.624617675Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 495.557454ms" Mar 7 01:13:20.627002 containerd[1638]: time="2026-03-07T01:13:20.626947131Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 504.620641ms" Mar 7 01:13:20.724345 containerd[1638]: time="2026-03-07T01:13:20.724028998Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:13:20.724345 containerd[1638]: time="2026-03-07T01:13:20.724062768Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:13:20.724345 containerd[1638]: time="2026-03-07T01:13:20.724069858Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:20.724345 containerd[1638]: time="2026-03-07T01:13:20.724128298Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:20.732677 containerd[1638]: time="2026-03-07T01:13:20.732481939Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:13:20.732677 containerd[1638]: time="2026-03-07T01:13:20.732529479Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:13:20.732677 containerd[1638]: time="2026-03-07T01:13:20.732547999Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:20.732677 containerd[1638]: time="2026-03-07T01:13:20.732624248Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:20.734771 containerd[1638]: time="2026-03-07T01:13:20.734614556Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:13:20.734771 containerd[1638]: time="2026-03-07T01:13:20.734648895Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:13:20.734771 containerd[1638]: time="2026-03-07T01:13:20.734658975Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:20.734771 containerd[1638]: time="2026-03-07T01:13:20.734718345Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:20.750068 kubelet[2369]: E0307 01:13:20.750036 2369 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://95.216.146.133:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-6-n-0c8881e772&limit=500&resourceVersion=0\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 7 01:13:20.800053 containerd[1638]: time="2026-03-07T01:13:20.800019381Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-6-n-0c8881e772,Uid:a160ce3374f2ce383338f0b710d12611,Namespace:kube-system,Attempt:0,} returns sandbox id \"95bb513489901d098904ca1c6fbe104c8a95276980dad00f918af6d1d85dc0b8\"" Mar 7 01:13:20.800954 containerd[1638]: time="2026-03-07T01:13:20.800908396Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-6-n-0c8881e772,Uid:4a90aea135cdd8a005d68e1fd7967f4d,Namespace:kube-system,Attempt:0,} returns sandbox id \"9a63b518a7686f622c2fd4f66058c82b795f3ef945f73b6933fd17807b362f0b\"" Mar 7 01:13:20.803210 containerd[1638]: time="2026-03-07T01:13:20.803171532Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-6-n-0c8881e772,Uid:cf82c8171bce44ee9aaf14eedd4a2d4c,Namespace:kube-system,Attempt:0,} returns sandbox id \"00c10846e3d6b3c18a74763e0e103d810d570a60b5bd8e0d2ca5c2a955c845a7\"" Mar 7 01:13:20.804758 containerd[1638]: time="2026-03-07T01:13:20.804741654Z" level=info msg="CreateContainer within sandbox \"95bb513489901d098904ca1c6fbe104c8a95276980dad00f918af6d1d85dc0b8\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 7 01:13:20.805981 containerd[1638]: time="2026-03-07T01:13:20.805965096Z" level=info msg="CreateContainer within sandbox \"9a63b518a7686f622c2fd4f66058c82b795f3ef945f73b6933fd17807b362f0b\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 7 01:13:20.806776 containerd[1638]: time="2026-03-07T01:13:20.806744361Z" level=info msg="CreateContainer within sandbox \"00c10846e3d6b3c18a74763e0e103d810d570a60b5bd8e0d2ca5c2a955c845a7\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 7 01:13:20.816711 kubelet[2369]: E0307 01:13:20.816344 2369 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://95.216.146.133:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 7 01:13:20.816764 containerd[1638]: time="2026-03-07T01:13:20.816635584Z" level=info msg="CreateContainer within sandbox \"95bb513489901d098904ca1c6fbe104c8a95276980dad00f918af6d1d85dc0b8\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"457a7ea62f0083e306d1f525084462df8f017aa3af4a00716124226e79599df7\"" Mar 7 01:13:20.817054 containerd[1638]: time="2026-03-07T01:13:20.817039081Z" level=info msg="StartContainer for \"457a7ea62f0083e306d1f525084462df8f017aa3af4a00716124226e79599df7\"" Mar 7 01:13:20.825000 containerd[1638]: time="2026-03-07T01:13:20.824962884Z" level=info msg="CreateContainer within sandbox \"00c10846e3d6b3c18a74763e0e103d810d570a60b5bd8e0d2ca5c2a955c845a7\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"cab76e9043845a9f8d584e7d51879f2931b4bee5fe16786c88660c637695e7b4\"" Mar 7 01:13:20.825294 containerd[1638]: time="2026-03-07T01:13:20.825265122Z" level=info msg="StartContainer for \"cab76e9043845a9f8d584e7d51879f2931b4bee5fe16786c88660c637695e7b4\"" Mar 7 01:13:20.826728 containerd[1638]: time="2026-03-07T01:13:20.826707784Z" level=info msg="CreateContainer within sandbox \"9a63b518a7686f622c2fd4f66058c82b795f3ef945f73b6933fd17807b362f0b\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"94247dfce2be4d6b909b5a17de58e02563d1a5396298d48b43afdae7e80ebc88\"" Mar 7 01:13:20.827032 containerd[1638]: time="2026-03-07T01:13:20.827007972Z" level=info msg="StartContainer for \"94247dfce2be4d6b909b5a17de58e02563d1a5396298d48b43afdae7e80ebc88\"" Mar 7 01:13:20.887563 kubelet[2369]: E0307 01:13:20.886208 2369 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://95.216.146.133:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 7 01:13:20.899618 containerd[1638]: time="2026-03-07T01:13:20.899572005Z" level=info msg="StartContainer for \"457a7ea62f0083e306d1f525084462df8f017aa3af4a00716124226e79599df7\" returns successfully" Mar 7 01:13:20.917431 containerd[1638]: time="2026-03-07T01:13:20.917397730Z" level=info msg="StartContainer for \"94247dfce2be4d6b909b5a17de58e02563d1a5396298d48b43afdae7e80ebc88\" returns successfully" Mar 7 01:13:20.923959 kubelet[2369]: E0307 01:13:20.923314 2369 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://95.216.146.133:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 95.216.146.133:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 7 01:13:20.938612 containerd[1638]: time="2026-03-07T01:13:20.938535265Z" level=info msg="StartContainer for \"cab76e9043845a9f8d584e7d51879f2931b4bee5fe16786c88660c637695e7b4\" returns successfully" Mar 7 01:13:21.239270 kubelet[2369]: I0307 01:13:21.239240 2369 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:21.714087 kubelet[2369]: E0307 01:13:21.713964 2369 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:21.715172 kubelet[2369]: E0307 01:13:21.715124 2369 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:21.715409 kubelet[2369]: E0307 01:13:21.715338 2369 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:22.364201 kubelet[2369]: E0307 01:13:22.364161 2369 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:22.563644 kubelet[2369]: I0307 01:13:22.563609 2369 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:22.563644 kubelet[2369]: E0307 01:13:22.563637 2369 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-4081-3-6-n-0c8881e772\": node \"ci-4081-3-6-n-0c8881e772\" not found" Mar 7 01:13:22.576323 kubelet[2369]: E0307 01:13:22.576280 2369 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-0c8881e772\" not found" Mar 7 01:13:22.676684 kubelet[2369]: E0307 01:13:22.676651 2369 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-0c8881e772\" not found" Mar 7 01:13:22.718380 kubelet[2369]: E0307 01:13:22.718342 2369 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:22.718874 kubelet[2369]: E0307 01:13:22.718705 2369 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-0c8881e772\" not found" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:22.776967 kubelet[2369]: E0307 01:13:22.776846 2369 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-0c8881e772\" not found" Mar 7 01:13:22.878075 kubelet[2369]: E0307 01:13:22.878016 2369 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-0c8881e772\" not found" Mar 7 01:13:22.979166 kubelet[2369]: E0307 01:13:22.978988 2369 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-0c8881e772\" not found" Mar 7 01:13:23.071738 kubelet[2369]: I0307 01:13:23.071646 2369 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:23.077016 kubelet[2369]: E0307 01:13:23.076995 2369 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-6-n-0c8881e772\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:23.077016 kubelet[2369]: I0307 01:13:23.077012 2369 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:23.077940 kubelet[2369]: E0307 01:13:23.077887 2369 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-0c8881e772\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:23.077940 kubelet[2369]: I0307 01:13:23.077902 2369 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:23.079256 kubelet[2369]: E0307 01:13:23.079216 2369 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:23.657962 kubelet[2369]: I0307 01:13:23.657916 2369 apiserver.go:52] "Watching apiserver" Mar 7 01:13:23.671837 kubelet[2369]: I0307 01:13:23.671689 2369 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 7 01:13:23.718813 kubelet[2369]: I0307 01:13:23.718764 2369 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:24.443809 systemd[1]: Reloading requested from client PID 2647 ('systemctl') (unit session-7.scope)... Mar 7 01:13:24.443840 systemd[1]: Reloading... Mar 7 01:13:24.547589 zram_generator::config[2693]: No configuration found. Mar 7 01:13:24.631925 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:13:24.696936 systemd[1]: Reloading finished in 252 ms. Mar 7 01:13:24.733612 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:13:24.749586 systemd[1]: kubelet.service: Deactivated successfully. Mar 7 01:13:24.749916 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:13:24.756777 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:13:24.924403 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:13:24.928023 (kubelet)[2747]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 7 01:13:24.959631 kubelet[2747]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 01:13:24.959631 kubelet[2747]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 7 01:13:24.959631 kubelet[2747]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 01:13:24.959631 kubelet[2747]: I0307 01:13:24.959574 2747 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 7 01:13:24.965705 kubelet[2747]: I0307 01:13:24.965667 2747 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 7 01:13:24.965705 kubelet[2747]: I0307 01:13:24.965698 2747 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 7 01:13:24.965886 kubelet[2747]: I0307 01:13:24.965861 2747 server.go:956] "Client rotation is on, will bootstrap in background" Mar 7 01:13:24.966818 kubelet[2747]: I0307 01:13:24.966802 2747 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 7 01:13:24.969906 kubelet[2747]: I0307 01:13:24.969758 2747 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 7 01:13:24.972021 kubelet[2747]: E0307 01:13:24.971999 2747 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 7 01:13:24.972064 kubelet[2747]: I0307 01:13:24.972018 2747 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 7 01:13:24.975263 kubelet[2747]: I0307 01:13:24.975242 2747 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 7 01:13:24.975683 kubelet[2747]: I0307 01:13:24.975658 2747 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 7 01:13:24.975772 kubelet[2747]: I0307 01:13:24.975674 2747 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-6-n-0c8881e772","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Mar 7 01:13:24.975772 kubelet[2747]: I0307 01:13:24.975769 2747 topology_manager.go:138] "Creating topology manager with none policy" Mar 7 01:13:24.975848 kubelet[2747]: I0307 01:13:24.975775 2747 container_manager_linux.go:303] "Creating device plugin manager" Mar 7 01:13:24.975848 kubelet[2747]: I0307 01:13:24.975811 2747 state_mem.go:36] "Initialized new in-memory state store" Mar 7 01:13:24.975984 kubelet[2747]: I0307 01:13:24.975965 2747 kubelet.go:480] "Attempting to sync node with API server" Mar 7 01:13:24.975984 kubelet[2747]: I0307 01:13:24.975977 2747 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 7 01:13:24.976018 kubelet[2747]: I0307 01:13:24.975996 2747 kubelet.go:386] "Adding apiserver pod source" Mar 7 01:13:24.978152 kubelet[2747]: I0307 01:13:24.977588 2747 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 7 01:13:24.981940 kubelet[2747]: I0307 01:13:24.981922 2747 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 7 01:13:24.982731 kubelet[2747]: I0307 01:13:24.982710 2747 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 7 01:13:24.986291 kubelet[2747]: I0307 01:13:24.986150 2747 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 7 01:13:24.986291 kubelet[2747]: I0307 01:13:24.986179 2747 server.go:1289] "Started kubelet" Mar 7 01:13:24.986291 kubelet[2747]: I0307 01:13:24.986229 2747 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 7 01:13:24.986811 kubelet[2747]: I0307 01:13:24.986794 2747 server.go:317] "Adding debug handlers to kubelet server" Mar 7 01:13:24.987054 kubelet[2747]: I0307 01:13:24.987022 2747 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 7 01:13:24.987257 kubelet[2747]: I0307 01:13:24.987243 2747 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 7 01:13:24.992941 kubelet[2747]: I0307 01:13:24.992837 2747 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 7 01:13:24.994033 kubelet[2747]: I0307 01:13:24.993572 2747 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 7 01:13:24.995010 kubelet[2747]: I0307 01:13:24.994817 2747 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 7 01:13:24.999565 kubelet[2747]: I0307 01:13:24.999257 2747 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 7 01:13:24.999565 kubelet[2747]: I0307 01:13:24.999354 2747 reconciler.go:26] "Reconciler: start to sync state" Mar 7 01:13:25.002217 kubelet[2747]: I0307 01:13:25.002204 2747 factory.go:223] Registration of the systemd container factory successfully Mar 7 01:13:25.002696 kubelet[2747]: I0307 01:13:25.002665 2747 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 7 01:13:25.004038 kubelet[2747]: I0307 01:13:25.004007 2747 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 7 01:13:25.004568 kubelet[2747]: I0307 01:13:25.004526 2747 factory.go:223] Registration of the containerd container factory successfully Mar 7 01:13:25.014406 kubelet[2747]: I0307 01:13:25.014391 2747 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 7 01:13:25.014462 kubelet[2747]: I0307 01:13:25.014456 2747 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 7 01:13:25.014512 kubelet[2747]: I0307 01:13:25.014505 2747 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 7 01:13:25.014578 kubelet[2747]: I0307 01:13:25.014571 2747 kubelet.go:2436] "Starting kubelet main sync loop" Mar 7 01:13:25.014671 kubelet[2747]: E0307 01:13:25.014657 2747 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 7 01:13:25.067611 kubelet[2747]: I0307 01:13:25.067579 2747 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 7 01:13:25.067611 kubelet[2747]: I0307 01:13:25.067594 2747 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 7 01:13:25.067611 kubelet[2747]: I0307 01:13:25.067608 2747 state_mem.go:36] "Initialized new in-memory state store" Mar 7 01:13:25.067750 kubelet[2747]: I0307 01:13:25.067711 2747 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 7 01:13:25.067750 kubelet[2747]: I0307 01:13:25.067718 2747 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 7 01:13:25.067750 kubelet[2747]: I0307 01:13:25.067730 2747 policy_none.go:49] "None policy: Start" Mar 7 01:13:25.067750 kubelet[2747]: I0307 01:13:25.067737 2747 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 7 01:13:25.067750 kubelet[2747]: I0307 01:13:25.067745 2747 state_mem.go:35] "Initializing new in-memory state store" Mar 7 01:13:25.067835 kubelet[2747]: I0307 01:13:25.067801 2747 state_mem.go:75] "Updated machine memory state" Mar 7 01:13:25.070276 kubelet[2747]: E0307 01:13:25.069043 2747 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 7 01:13:25.070276 kubelet[2747]: I0307 01:13:25.069173 2747 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 7 01:13:25.070276 kubelet[2747]: I0307 01:13:25.069182 2747 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 7 01:13:25.070276 kubelet[2747]: I0307 01:13:25.070049 2747 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 7 01:13:25.072193 kubelet[2747]: E0307 01:13:25.072017 2747 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 7 01:13:25.115762 kubelet[2747]: I0307 01:13:25.115714 2747 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.115999 kubelet[2747]: I0307 01:13:25.115988 2747 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.116123 kubelet[2747]: I0307 01:13:25.116103 2747 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.122912 kubelet[2747]: E0307 01:13:25.122833 2747 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-6-n-0c8881e772\" already exists" pod="kube-system/kube-scheduler-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.176436 kubelet[2747]: I0307 01:13:25.176378 2747 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.184472 kubelet[2747]: I0307 01:13:25.184423 2747 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.184631 kubelet[2747]: I0307 01:13:25.184512 2747 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.200872 kubelet[2747]: I0307 01:13:25.200835 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.201050 kubelet[2747]: I0307 01:13:25.200898 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.201050 kubelet[2747]: I0307 01:13:25.200924 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.201050 kubelet[2747]: I0307 01:13:25.200948 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/cf82c8171bce44ee9aaf14eedd4a2d4c-kubeconfig\") pod \"kube-scheduler-ci-4081-3-6-n-0c8881e772\" (UID: \"cf82c8171bce44ee9aaf14eedd4a2d4c\") " pod="kube-system/kube-scheduler-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.201050 kubelet[2747]: I0307 01:13:25.200968 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4a90aea135cdd8a005d68e1fd7967f4d-k8s-certs\") pod \"kube-apiserver-ci-4081-3-6-n-0c8881e772\" (UID: \"4a90aea135cdd8a005d68e1fd7967f4d\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.201050 kubelet[2747]: I0307 01:13:25.200985 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-ca-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.201210 kubelet[2747]: I0307 01:13:25.201003 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a160ce3374f2ce383338f0b710d12611-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-6-n-0c8881e772\" (UID: \"a160ce3374f2ce383338f0b710d12611\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.201210 kubelet[2747]: I0307 01:13:25.201022 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4a90aea135cdd8a005d68e1fd7967f4d-ca-certs\") pod \"kube-apiserver-ci-4081-3-6-n-0c8881e772\" (UID: \"4a90aea135cdd8a005d68e1fd7967f4d\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.201210 kubelet[2747]: I0307 01:13:25.201041 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4a90aea135cdd8a005d68e1fd7967f4d-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-6-n-0c8881e772\" (UID: \"4a90aea135cdd8a005d68e1fd7967f4d\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:25.979087 kubelet[2747]: I0307 01:13:25.979039 2747 apiserver.go:52] "Watching apiserver" Mar 7 01:13:26.000129 kubelet[2747]: I0307 01:13:26.000089 2747 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 7 01:13:26.046303 kubelet[2747]: I0307 01:13:26.044747 2747 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:26.048842 kubelet[2747]: E0307 01:13:26.048742 2747 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-0c8881e772\" already exists" pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" Mar 7 01:13:26.066528 kubelet[2747]: I0307 01:13:26.066474 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" podStartSLOduration=1.066448338 podStartE2EDuration="1.066448338s" podCreationTimestamp="2026-03-07 01:13:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:13:26.065703711 +0000 UTC m=+1.132615428" watchObservedRunningTime="2026-03-07 01:13:26.066448338 +0000 UTC m=+1.133360065" Mar 7 01:13:26.079952 kubelet[2747]: I0307 01:13:26.079904 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-0c8881e772" podStartSLOduration=1.079893716 podStartE2EDuration="1.079893716s" podCreationTimestamp="2026-03-07 01:13:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:13:26.071896167 +0000 UTC m=+1.138807894" watchObservedRunningTime="2026-03-07 01:13:26.079893716 +0000 UTC m=+1.146805443" Mar 7 01:13:26.079952 kubelet[2747]: I0307 01:13:26.079953 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081-3-6-n-0c8881e772" podStartSLOduration=3.079951706 podStartE2EDuration="3.079951706s" podCreationTimestamp="2026-03-07 01:13:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:13:26.079499407 +0000 UTC m=+1.146411124" watchObservedRunningTime="2026-03-07 01:13:26.079951706 +0000 UTC m=+1.146863423" Mar 7 01:13:29.566874 kubelet[2747]: I0307 01:13:29.566837 2747 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 7 01:13:29.567572 containerd[1638]: time="2026-03-07T01:13:29.567396008Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 7 01:13:29.567819 kubelet[2747]: I0307 01:13:29.567555 2747 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 7 01:13:30.635156 kubelet[2747]: I0307 01:13:30.632581 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/14c2224d-6ba6-4fa5-88e4-27e1a0395fcf-kube-proxy\") pod \"kube-proxy-78xzl\" (UID: \"14c2224d-6ba6-4fa5-88e4-27e1a0395fcf\") " pod="kube-system/kube-proxy-78xzl" Mar 7 01:13:30.635156 kubelet[2747]: I0307 01:13:30.632609 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/14c2224d-6ba6-4fa5-88e4-27e1a0395fcf-lib-modules\") pod \"kube-proxy-78xzl\" (UID: \"14c2224d-6ba6-4fa5-88e4-27e1a0395fcf\") " pod="kube-system/kube-proxy-78xzl" Mar 7 01:13:30.635156 kubelet[2747]: I0307 01:13:30.632623 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-772gj\" (UniqueName: \"kubernetes.io/projected/14c2224d-6ba6-4fa5-88e4-27e1a0395fcf-kube-api-access-772gj\") pod \"kube-proxy-78xzl\" (UID: \"14c2224d-6ba6-4fa5-88e4-27e1a0395fcf\") " pod="kube-system/kube-proxy-78xzl" Mar 7 01:13:30.635156 kubelet[2747]: I0307 01:13:30.632638 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/14c2224d-6ba6-4fa5-88e4-27e1a0395fcf-xtables-lock\") pod \"kube-proxy-78xzl\" (UID: \"14c2224d-6ba6-4fa5-88e4-27e1a0395fcf\") " pod="kube-system/kube-proxy-78xzl" Mar 7 01:13:30.733729 kubelet[2747]: I0307 01:13:30.733692 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sptg\" (UniqueName: \"kubernetes.io/projected/7e4a93f4-9c0e-44f9-bb3a-dcba317b694e-kube-api-access-2sptg\") pod \"tigera-operator-6bf85f8dd-ljzps\" (UID: \"7e4a93f4-9c0e-44f9-bb3a-dcba317b694e\") " pod="tigera-operator/tigera-operator-6bf85f8dd-ljzps" Mar 7 01:13:30.733928 kubelet[2747]: I0307 01:13:30.733911 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/7e4a93f4-9c0e-44f9-bb3a-dcba317b694e-var-lib-calico\") pod \"tigera-operator-6bf85f8dd-ljzps\" (UID: \"7e4a93f4-9c0e-44f9-bb3a-dcba317b694e\") " pod="tigera-operator/tigera-operator-6bf85f8dd-ljzps" Mar 7 01:13:30.860707 containerd[1638]: time="2026-03-07T01:13:30.860660405Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-78xzl,Uid:14c2224d-6ba6-4fa5-88e4-27e1a0395fcf,Namespace:kube-system,Attempt:0,}" Mar 7 01:13:30.883730 containerd[1638]: time="2026-03-07T01:13:30.883629199Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:13:30.884512 containerd[1638]: time="2026-03-07T01:13:30.884391127Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:13:30.884695 containerd[1638]: time="2026-03-07T01:13:30.884493987Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:30.885389 containerd[1638]: time="2026-03-07T01:13:30.885191594Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:30.924817 containerd[1638]: time="2026-03-07T01:13:30.924769959Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-78xzl,Uid:14c2224d-6ba6-4fa5-88e4-27e1a0395fcf,Namespace:kube-system,Attempt:0,} returns sandbox id \"a82c5b63456392d28cef8499674a3934e0b7eac5c57fc37828dd27500b7c6710\"" Mar 7 01:13:30.929765 containerd[1638]: time="2026-03-07T01:13:30.929714926Z" level=info msg="CreateContainer within sandbox \"a82c5b63456392d28cef8499674a3934e0b7eac5c57fc37828dd27500b7c6710\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 7 01:13:30.934365 containerd[1638]: time="2026-03-07T01:13:30.934250132Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6bf85f8dd-ljzps,Uid:7e4a93f4-9c0e-44f9-bb3a-dcba317b694e,Namespace:tigera-operator,Attempt:0,}" Mar 7 01:13:30.945151 containerd[1638]: time="2026-03-07T01:13:30.945032171Z" level=info msg="CreateContainer within sandbox \"a82c5b63456392d28cef8499674a3934e0b7eac5c57fc37828dd27500b7c6710\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"efe1ed457e27eb3f23e617e58f9b7528ec68fcd085398e99f523e0d4d2735d9a\"" Mar 7 01:13:30.947491 containerd[1638]: time="2026-03-07T01:13:30.946684327Z" level=info msg="StartContainer for \"efe1ed457e27eb3f23e617e58f9b7528ec68fcd085398e99f523e0d4d2735d9a\"" Mar 7 01:13:30.969359 containerd[1638]: time="2026-03-07T01:13:30.968488314Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:13:30.969359 containerd[1638]: time="2026-03-07T01:13:30.968592054Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:13:30.969359 containerd[1638]: time="2026-03-07T01:13:30.968601234Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:30.969359 containerd[1638]: time="2026-03-07T01:13:30.968667533Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:31.012029 containerd[1638]: time="2026-03-07T01:13:31.011994110Z" level=info msg="StartContainer for \"efe1ed457e27eb3f23e617e58f9b7528ec68fcd085398e99f523e0d4d2735d9a\" returns successfully" Mar 7 01:13:31.017712 containerd[1638]: time="2026-03-07T01:13:31.017685495Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6bf85f8dd-ljzps,Uid:7e4a93f4-9c0e-44f9-bb3a-dcba317b694e,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"5bf1d14efe9b97f40fb6b10ebe472dace799d77f722cb43428ea999cb644f370\"" Mar 7 01:13:31.020183 containerd[1638]: time="2026-03-07T01:13:31.020145228Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\"" Mar 7 01:13:32.647607 update_engine[1622]: I20260307 01:13:32.647465 1622 update_attempter.cc:509] Updating boot flags... Mar 7 01:13:32.716715 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 34 scanned by (udev-worker) (3052) Mar 7 01:13:32.784351 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 34 scanned by (udev-worker) (2950) Mar 7 01:13:32.948359 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount761805372.mount: Deactivated successfully. Mar 7 01:13:33.480558 containerd[1638]: time="2026-03-07T01:13:33.480494020Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:33.481558 containerd[1638]: time="2026-03-07T01:13:33.481503588Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.7: active requests=0, bytes read=40846156" Mar 7 01:13:33.482583 containerd[1638]: time="2026-03-07T01:13:33.482551106Z" level=info msg="ImageCreate event name:\"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:33.484460 containerd[1638]: time="2026-03-07T01:13:33.484439171Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:33.485034 containerd[1638]: time="2026-03-07T01:13:33.484946980Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.7\" with image id \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\", repo tag \"quay.io/tigera/operator:v1.40.7\", repo digest \"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\", size \"40842151\" in 2.464774842s" Mar 7 01:13:33.485034 containerd[1638]: time="2026-03-07T01:13:33.484969670Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\" returns image reference \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\"" Mar 7 01:13:33.488409 containerd[1638]: time="2026-03-07T01:13:33.488381272Z" level=info msg="CreateContainer within sandbox \"5bf1d14efe9b97f40fb6b10ebe472dace799d77f722cb43428ea999cb644f370\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Mar 7 01:13:33.501483 containerd[1638]: time="2026-03-07T01:13:33.501454242Z" level=info msg="CreateContainer within sandbox \"5bf1d14efe9b97f40fb6b10ebe472dace799d77f722cb43428ea999cb644f370\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe\"" Mar 7 01:13:33.501811 containerd[1638]: time="2026-03-07T01:13:33.501792111Z" level=info msg="StartContainer for \"3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe\"" Mar 7 01:13:33.544487 containerd[1638]: time="2026-03-07T01:13:33.543694794Z" level=info msg="StartContainer for \"3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe\" returns successfully" Mar 7 01:13:34.075532 kubelet[2747]: I0307 01:13:34.075258 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-78xzl" podStartSLOduration=4.075236381 podStartE2EDuration="4.075236381s" podCreationTimestamp="2026-03-07 01:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:13:31.070647093 +0000 UTC m=+6.137558810" watchObservedRunningTime="2026-03-07 01:13:34.075236381 +0000 UTC m=+9.142148138" Mar 7 01:13:38.465726 sudo[1865]: pam_unix(sudo:session): session closed for user root Mar 7 01:13:38.588887 sshd[1861]: pam_unix(sshd:session): session closed for user core Mar 7 01:13:38.591650 systemd[1]: sshd@6-95.216.146.133:22-4.153.228.146:51270.service: Deactivated successfully. Mar 7 01:13:38.594806 systemd[1]: session-7.scope: Deactivated successfully. Mar 7 01:13:38.595514 systemd-logind[1620]: Session 7 logged out. Waiting for processes to exit. Mar 7 01:13:38.596967 systemd-logind[1620]: Removed session 7. Mar 7 01:13:38.889626 kubelet[2747]: I0307 01:13:38.889519 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6bf85f8dd-ljzps" podStartSLOduration=6.423134779 podStartE2EDuration="8.889507197s" podCreationTimestamp="2026-03-07 01:13:30 +0000 UTC" firstStartedPulling="2026-03-07 01:13:31.019125511 +0000 UTC m=+6.086037228" lastFinishedPulling="2026-03-07 01:13:33.485497929 +0000 UTC m=+8.552409646" observedRunningTime="2026-03-07 01:13:34.07550447 +0000 UTC m=+9.142416227" watchObservedRunningTime="2026-03-07 01:13:38.889507197 +0000 UTC m=+13.956418924" Mar 7 01:13:40.494344 kubelet[2747]: I0307 01:13:40.493773 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2144b56e-33de-4010-bbdd-c43fd2475741-tigera-ca-bundle\") pod \"calico-typha-59c77dc479-mgw2d\" (UID: \"2144b56e-33de-4010-bbdd-c43fd2475741\") " pod="calico-system/calico-typha-59c77dc479-mgw2d" Mar 7 01:13:40.494344 kubelet[2747]: I0307 01:13:40.493806 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc5xx\" (UniqueName: \"kubernetes.io/projected/2144b56e-33de-4010-bbdd-c43fd2475741-kube-api-access-zc5xx\") pod \"calico-typha-59c77dc479-mgw2d\" (UID: \"2144b56e-33de-4010-bbdd-c43fd2475741\") " pod="calico-system/calico-typha-59c77dc479-mgw2d" Mar 7 01:13:40.494344 kubelet[2747]: I0307 01:13:40.493818 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/2144b56e-33de-4010-bbdd-c43fd2475741-typha-certs\") pod \"calico-typha-59c77dc479-mgw2d\" (UID: \"2144b56e-33de-4010-bbdd-c43fd2475741\") " pod="calico-system/calico-typha-59c77dc479-mgw2d" Mar 7 01:13:40.624904 kubelet[2747]: E0307 01:13:40.624016 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:40.695385 kubelet[2747]: I0307 01:13:40.695346 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/eb326673-d2f8-44d7-8c06-d4f44d6af381-node-certs\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695385 kubelet[2747]: I0307 01:13:40.695379 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-policysync\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695385 kubelet[2747]: I0307 01:13:40.695392 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-cni-bin-dir\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695526 kubelet[2747]: I0307 01:13:40.695403 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-cni-log-dir\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695526 kubelet[2747]: I0307 01:13:40.695415 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-var-lib-calico\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695526 kubelet[2747]: I0307 01:13:40.695428 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-cni-net-dir\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695526 kubelet[2747]: I0307 01:13:40.695439 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-lib-modules\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695526 kubelet[2747]: I0307 01:13:40.695450 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-nodeproc\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695633 kubelet[2747]: I0307 01:13:40.695460 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-sys-fs\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695633 kubelet[2747]: I0307 01:13:40.695471 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-var-run-calico\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695633 kubelet[2747]: I0307 01:13:40.695481 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-bpffs\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695633 kubelet[2747]: I0307 01:13:40.695491 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-xtables-lock\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695633 kubelet[2747]: I0307 01:13:40.695502 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjblr\" (UniqueName: \"kubernetes.io/projected/eb326673-d2f8-44d7-8c06-d4f44d6af381-kube-api-access-kjblr\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695709 kubelet[2747]: I0307 01:13:40.695513 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/eb326673-d2f8-44d7-8c06-d4f44d6af381-flexvol-driver-host\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.695709 kubelet[2747]: I0307 01:13:40.695523 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb326673-d2f8-44d7-8c06-d4f44d6af381-tigera-ca-bundle\") pod \"calico-node-lscfk\" (UID: \"eb326673-d2f8-44d7-8c06-d4f44d6af381\") " pod="calico-system/calico-node-lscfk" Mar 7 01:13:40.760507 containerd[1638]: time="2026-03-07T01:13:40.760423271Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-59c77dc479-mgw2d,Uid:2144b56e-33de-4010-bbdd-c43fd2475741,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:40.782248 containerd[1638]: time="2026-03-07T01:13:40.781979092Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:13:40.782248 containerd[1638]: time="2026-03-07T01:13:40.782025132Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:13:40.782248 containerd[1638]: time="2026-03-07T01:13:40.782046812Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:40.782248 containerd[1638]: time="2026-03-07T01:13:40.782113962Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:40.797756 kubelet[2747]: I0307 01:13:40.796225 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7c7g\" (UniqueName: \"kubernetes.io/projected/1fd3aff8-915e-4a01-9844-abd28ee38035-kube-api-access-s7c7g\") pod \"csi-node-driver-xbz8b\" (UID: \"1fd3aff8-915e-4a01-9844-abd28ee38035\") " pod="calico-system/csi-node-driver-xbz8b" Mar 7 01:13:40.797830 kubelet[2747]: E0307 01:13:40.797808 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.797830 kubelet[2747]: W0307 01:13:40.797827 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.797870 kubelet[2747]: E0307 01:13:40.797839 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.798523 kubelet[2747]: E0307 01:13:40.798071 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.798523 kubelet[2747]: W0307 01:13:40.798088 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.798523 kubelet[2747]: E0307 01:13:40.798095 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.798523 kubelet[2747]: I0307 01:13:40.798110 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1fd3aff8-915e-4a01-9844-abd28ee38035-registration-dir\") pod \"csi-node-driver-xbz8b\" (UID: \"1fd3aff8-915e-4a01-9844-abd28ee38035\") " pod="calico-system/csi-node-driver-xbz8b" Mar 7 01:13:40.798523 kubelet[2747]: E0307 01:13:40.798467 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.798523 kubelet[2747]: W0307 01:13:40.798474 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.798523 kubelet[2747]: E0307 01:13:40.798480 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.801285 kubelet[2747]: E0307 01:13:40.801267 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.801339 kubelet[2747]: W0307 01:13:40.801328 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.801393 kubelet[2747]: E0307 01:13:40.801384 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.801729 kubelet[2747]: I0307 01:13:40.801711 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/1fd3aff8-915e-4a01-9844-abd28ee38035-varrun\") pod \"csi-node-driver-xbz8b\" (UID: \"1fd3aff8-915e-4a01-9844-abd28ee38035\") " pod="calico-system/csi-node-driver-xbz8b" Mar 7 01:13:40.801810 kubelet[2747]: E0307 01:13:40.801801 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.801840 kubelet[2747]: W0307 01:13:40.801832 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.801890 kubelet[2747]: E0307 01:13:40.801883 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.802158 kubelet[2747]: E0307 01:13:40.802150 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.802203 kubelet[2747]: W0307 01:13:40.802196 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.802239 kubelet[2747]: E0307 01:13:40.802231 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.802485 kubelet[2747]: E0307 01:13:40.802476 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.802528 kubelet[2747]: W0307 01:13:40.802521 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.802602 kubelet[2747]: E0307 01:13:40.802594 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.802849 kubelet[2747]: E0307 01:13:40.802841 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.802891 kubelet[2747]: W0307 01:13:40.802885 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.802944 kubelet[2747]: E0307 01:13:40.802937 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.803161 kubelet[2747]: E0307 01:13:40.803154 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.803197 kubelet[2747]: W0307 01:13:40.803190 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.803225 kubelet[2747]: E0307 01:13:40.803219 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.803431 kubelet[2747]: E0307 01:13:40.803424 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.803465 kubelet[2747]: W0307 01:13:40.803459 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.803507 kubelet[2747]: E0307 01:13:40.803500 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.803731 kubelet[2747]: E0307 01:13:40.803723 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.803822 kubelet[2747]: W0307 01:13:40.803773 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.803822 kubelet[2747]: E0307 01:13:40.803782 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.804072 kubelet[2747]: E0307 01:13:40.804019 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.804072 kubelet[2747]: W0307 01:13:40.804026 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.804072 kubelet[2747]: E0307 01:13:40.804033 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.804351 kubelet[2747]: E0307 01:13:40.804275 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.804351 kubelet[2747]: W0307 01:13:40.804282 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.804351 kubelet[2747]: E0307 01:13:40.804288 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.804615 kubelet[2747]: E0307 01:13:40.804525 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.804615 kubelet[2747]: W0307 01:13:40.804533 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.804615 kubelet[2747]: E0307 01:13:40.804559 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.804839 kubelet[2747]: E0307 01:13:40.804832 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.805065 kubelet[2747]: W0307 01:13:40.804873 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.805065 kubelet[2747]: E0307 01:13:40.804882 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.805177 kubelet[2747]: E0307 01:13:40.805159 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.805177 kubelet[2747]: W0307 01:13:40.805172 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.805218 kubelet[2747]: E0307 01:13:40.805182 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.805969 kubelet[2747]: E0307 01:13:40.805855 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.805969 kubelet[2747]: W0307 01:13:40.805865 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.805969 kubelet[2747]: E0307 01:13:40.805873 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.806294 kubelet[2747]: E0307 01:13:40.806286 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.806529 kubelet[2747]: W0307 01:13:40.806325 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.806529 kubelet[2747]: E0307 01:13:40.806333 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.807396 kubelet[2747]: E0307 01:13:40.807387 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.807464 kubelet[2747]: W0307 01:13:40.807456 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.807499 kubelet[2747]: E0307 01:13:40.807491 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.808473 kubelet[2747]: E0307 01:13:40.808463 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.808609 kubelet[2747]: W0307 01:13:40.808600 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.808952 kubelet[2747]: E0307 01:13:40.808670 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.811172 kubelet[2747]: E0307 01:13:40.811139 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.811172 kubelet[2747]: W0307 01:13:40.811149 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.811172 kubelet[2747]: E0307 01:13:40.811160 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.812632 kubelet[2747]: E0307 01:13:40.812622 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.812751 kubelet[2747]: W0307 01:13:40.812669 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.812751 kubelet[2747]: E0307 01:13:40.812682 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.813024 kubelet[2747]: E0307 01:13:40.812950 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.813395 kubelet[2747]: W0307 01:13:40.813247 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.813395 kubelet[2747]: E0307 01:13:40.813260 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.814174 kubelet[2747]: E0307 01:13:40.813958 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.814174 kubelet[2747]: W0307 01:13:40.813969 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.814174 kubelet[2747]: E0307 01:13:40.813982 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.814174 kubelet[2747]: E0307 01:13:40.814150 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.814174 kubelet[2747]: W0307 01:13:40.814156 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.814174 kubelet[2747]: E0307 01:13:40.814162 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.814672 kubelet[2747]: E0307 01:13:40.814373 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.814672 kubelet[2747]: W0307 01:13:40.814382 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.814672 kubelet[2747]: E0307 01:13:40.814388 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.814672 kubelet[2747]: E0307 01:13:40.814625 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.814672 kubelet[2747]: W0307 01:13:40.814632 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.814672 kubelet[2747]: E0307 01:13:40.814638 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.815233 kubelet[2747]: E0307 01:13:40.815185 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.815233 kubelet[2747]: W0307 01:13:40.815207 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.815233 kubelet[2747]: E0307 01:13:40.815214 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.815464 kubelet[2747]: E0307 01:13:40.815432 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.815464 kubelet[2747]: W0307 01:13:40.815442 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.815464 kubelet[2747]: E0307 01:13:40.815451 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.815785 kubelet[2747]: E0307 01:13:40.815734 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.815785 kubelet[2747]: W0307 01:13:40.815742 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.815785 kubelet[2747]: E0307 01:13:40.815749 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.816216 kubelet[2747]: E0307 01:13:40.816127 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.816216 kubelet[2747]: W0307 01:13:40.816135 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.816216 kubelet[2747]: E0307 01:13:40.816142 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.816482 kubelet[2747]: E0307 01:13:40.816474 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.816519 kubelet[2747]: W0307 01:13:40.816512 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.816576 kubelet[2747]: E0307 01:13:40.816569 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.816834 kubelet[2747]: E0307 01:13:40.816811 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.816834 kubelet[2747]: W0307 01:13:40.816819 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.816834 kubelet[2747]: E0307 01:13:40.816825 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.817209 kubelet[2747]: E0307 01:13:40.817120 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.817209 kubelet[2747]: W0307 01:13:40.817127 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.817209 kubelet[2747]: E0307 01:13:40.817133 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.817209 kubelet[2747]: I0307 01:13:40.817158 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1fd3aff8-915e-4a01-9844-abd28ee38035-kubelet-dir\") pod \"csi-node-driver-xbz8b\" (UID: \"1fd3aff8-915e-4a01-9844-abd28ee38035\") " pod="calico-system/csi-node-driver-xbz8b" Mar 7 01:13:40.817488 kubelet[2747]: E0307 01:13:40.817463 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.817488 kubelet[2747]: W0307 01:13:40.817472 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.817488 kubelet[2747]: E0307 01:13:40.817479 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.817859 kubelet[2747]: E0307 01:13:40.817787 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.817859 kubelet[2747]: W0307 01:13:40.817812 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.817859 kubelet[2747]: E0307 01:13:40.817819 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.818273 kubelet[2747]: E0307 01:13:40.818194 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.818273 kubelet[2747]: W0307 01:13:40.818202 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.818273 kubelet[2747]: E0307 01:13:40.818210 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.818569 kubelet[2747]: E0307 01:13:40.818534 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.818640 kubelet[2747]: W0307 01:13:40.818610 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.818640 kubelet[2747]: E0307 01:13:40.818617 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.818983 kubelet[2747]: E0307 01:13:40.818926 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.818983 kubelet[2747]: W0307 01:13:40.818938 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.818983 kubelet[2747]: E0307 01:13:40.818944 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.819289 kubelet[2747]: E0307 01:13:40.819265 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.819289 kubelet[2747]: W0307 01:13:40.819274 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.819289 kubelet[2747]: E0307 01:13:40.819281 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.819627 kubelet[2747]: E0307 01:13:40.819568 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.819627 kubelet[2747]: W0307 01:13:40.819576 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.819627 kubelet[2747]: E0307 01:13:40.819582 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.819953 kubelet[2747]: E0307 01:13:40.819886 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.819953 kubelet[2747]: W0307 01:13:40.819894 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.819953 kubelet[2747]: E0307 01:13:40.819910 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.820268 kubelet[2747]: E0307 01:13:40.820185 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.820268 kubelet[2747]: W0307 01:13:40.820193 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.820268 kubelet[2747]: E0307 01:13:40.820200 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.820408 kubelet[2747]: E0307 01:13:40.820388 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.820455 kubelet[2747]: W0307 01:13:40.820395 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.820455 kubelet[2747]: E0307 01:13:40.820447 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.820790 kubelet[2747]: E0307 01:13:40.820759 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.820790 kubelet[2747]: W0307 01:13:40.820767 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.820790 kubelet[2747]: E0307 01:13:40.820773 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.821047 kubelet[2747]: E0307 01:13:40.821037 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.821095 kubelet[2747]: W0307 01:13:40.821078 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.821095 kubelet[2747]: E0307 01:13:40.821087 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.821390 kubelet[2747]: E0307 01:13:40.821317 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.821390 kubelet[2747]: W0307 01:13:40.821324 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.821390 kubelet[2747]: E0307 01:13:40.821331 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.821522 kubelet[2747]: E0307 01:13:40.821515 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.821584 kubelet[2747]: W0307 01:13:40.821566 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.821584 kubelet[2747]: E0307 01:13:40.821576 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.821847 kubelet[2747]: E0307 01:13:40.821816 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.821847 kubelet[2747]: W0307 01:13:40.821823 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.821847 kubelet[2747]: E0307 01:13:40.821830 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.822104 kubelet[2747]: E0307 01:13:40.822096 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.822146 kubelet[2747]: W0307 01:13:40.822139 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.822177 kubelet[2747]: E0307 01:13:40.822168 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.822415 kubelet[2747]: E0307 01:13:40.822392 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.822415 kubelet[2747]: W0307 01:13:40.822399 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.822415 kubelet[2747]: E0307 01:13:40.822406 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.822757 kubelet[2747]: E0307 01:13:40.822698 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.822757 kubelet[2747]: W0307 01:13:40.822706 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.822757 kubelet[2747]: E0307 01:13:40.822714 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.822979 kubelet[2747]: E0307 01:13:40.822971 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.823088 kubelet[2747]: W0307 01:13:40.823011 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.823088 kubelet[2747]: E0307 01:13:40.823019 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.823210 kubelet[2747]: E0307 01:13:40.823203 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.823246 kubelet[2747]: W0307 01:13:40.823239 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.823279 kubelet[2747]: E0307 01:13:40.823272 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.823631 kubelet[2747]: E0307 01:13:40.823524 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.823631 kubelet[2747]: W0307 01:13:40.823532 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.823631 kubelet[2747]: E0307 01:13:40.823565 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.823777 kubelet[2747]: E0307 01:13:40.823769 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.823810 kubelet[2747]: W0307 01:13:40.823803 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.823837 kubelet[2747]: E0307 01:13:40.823831 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.824113 kubelet[2747]: E0307 01:13:40.824105 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.824167 kubelet[2747]: W0307 01:13:40.824149 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.824167 kubelet[2747]: E0307 01:13:40.824158 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.824523 kubelet[2747]: E0307 01:13:40.824485 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.824523 kubelet[2747]: W0307 01:13:40.824493 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.824523 kubelet[2747]: E0307 01:13:40.824514 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.824826 kubelet[2747]: E0307 01:13:40.824802 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.824826 kubelet[2747]: W0307 01:13:40.824810 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.824826 kubelet[2747]: E0307 01:13:40.824817 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.825144 kubelet[2747]: E0307 01:13:40.825085 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.825144 kubelet[2747]: W0307 01:13:40.825093 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.825144 kubelet[2747]: E0307 01:13:40.825100 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.825381 kubelet[2747]: E0307 01:13:40.825359 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.825381 kubelet[2747]: W0307 01:13:40.825366 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.825381 kubelet[2747]: E0307 01:13:40.825373 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.825779 kubelet[2747]: E0307 01:13:40.825721 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.825779 kubelet[2747]: W0307 01:13:40.825728 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.825779 kubelet[2747]: E0307 01:13:40.825735 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.826085 kubelet[2747]: E0307 01:13:40.826014 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.826085 kubelet[2747]: W0307 01:13:40.826022 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.826085 kubelet[2747]: E0307 01:13:40.826029 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.826285 kubelet[2747]: E0307 01:13:40.826278 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.826328 kubelet[2747]: W0307 01:13:40.826312 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.826328 kubelet[2747]: E0307 01:13:40.826320 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.826677 kubelet[2747]: E0307 01:13:40.826605 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.826677 kubelet[2747]: W0307 01:13:40.826612 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.826677 kubelet[2747]: E0307 01:13:40.826618 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.826876 kubelet[2747]: E0307 01:13:40.826869 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.826975 kubelet[2747]: W0307 01:13:40.826923 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.826975 kubelet[2747]: E0307 01:13:40.826932 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.827263 kubelet[2747]: E0307 01:13:40.827225 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.827263 kubelet[2747]: W0307 01:13:40.827233 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.827263 kubelet[2747]: E0307 01:13:40.827240 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.827668 kubelet[2747]: E0307 01:13:40.827608 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.827668 kubelet[2747]: W0307 01:13:40.827617 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.827668 kubelet[2747]: E0307 01:13:40.827623 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.828052 kubelet[2747]: E0307 01:13:40.828002 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.828052 kubelet[2747]: W0307 01:13:40.828010 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.828052 kubelet[2747]: E0307 01:13:40.828017 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.828415 kubelet[2747]: E0307 01:13:40.828362 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.828415 kubelet[2747]: W0307 01:13:40.828370 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.828415 kubelet[2747]: E0307 01:13:40.828377 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.828850 kubelet[2747]: E0307 01:13:40.828778 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.828850 kubelet[2747]: W0307 01:13:40.828786 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.828850 kubelet[2747]: E0307 01:13:40.828793 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.829121 kubelet[2747]: E0307 01:13:40.829063 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.829121 kubelet[2747]: W0307 01:13:40.829070 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.829121 kubelet[2747]: E0307 01:13:40.829077 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.829248 kubelet[2747]: I0307 01:13:40.829206 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1fd3aff8-915e-4a01-9844-abd28ee38035-socket-dir\") pod \"csi-node-driver-xbz8b\" (UID: \"1fd3aff8-915e-4a01-9844-abd28ee38035\") " pod="calico-system/csi-node-driver-xbz8b" Mar 7 01:13:40.829747 kubelet[2747]: E0307 01:13:40.829668 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.829747 kubelet[2747]: W0307 01:13:40.829675 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.829747 kubelet[2747]: E0307 01:13:40.829682 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.831174 kubelet[2747]: E0307 01:13:40.831165 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.831238 kubelet[2747]: W0307 01:13:40.831208 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.831238 kubelet[2747]: E0307 01:13:40.831218 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.831740 kubelet[2747]: E0307 01:13:40.831485 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.831740 kubelet[2747]: W0307 01:13:40.831493 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.831740 kubelet[2747]: E0307 01:13:40.831499 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.832068 kubelet[2747]: E0307 01:13:40.831985 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.832068 kubelet[2747]: W0307 01:13:40.831993 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.832068 kubelet[2747]: E0307 01:13:40.832000 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.832567 kubelet[2747]: E0307 01:13:40.832512 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.832567 kubelet[2747]: W0307 01:13:40.832521 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.832567 kubelet[2747]: E0307 01:13:40.832528 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.833455 kubelet[2747]: E0307 01:13:40.833437 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.833521 kubelet[2747]: W0307 01:13:40.833500 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.833521 kubelet[2747]: E0307 01:13:40.833511 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.833817 kubelet[2747]: E0307 01:13:40.833809 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.834024 kubelet[2747]: W0307 01:13:40.833858 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.834432 kubelet[2747]: E0307 01:13:40.834170 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.835005 kubelet[2747]: E0307 01:13:40.834921 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.835005 kubelet[2747]: W0307 01:13:40.834929 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.835005 kubelet[2747]: E0307 01:13:40.834937 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.835160 kubelet[2747]: E0307 01:13:40.835152 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.835203 kubelet[2747]: W0307 01:13:40.835184 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.835203 kubelet[2747]: E0307 01:13:40.835194 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.835531 kubelet[2747]: E0307 01:13:40.835469 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.835531 kubelet[2747]: W0307 01:13:40.835478 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.835531 kubelet[2747]: E0307 01:13:40.835484 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.835865 kubelet[2747]: E0307 01:13:40.835804 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.835865 kubelet[2747]: W0307 01:13:40.835811 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.835865 kubelet[2747]: E0307 01:13:40.835819 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.836194 kubelet[2747]: E0307 01:13:40.836086 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.836194 kubelet[2747]: W0307 01:13:40.836094 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.836194 kubelet[2747]: E0307 01:13:40.836100 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.838621 containerd[1638]: time="2026-03-07T01:13:40.838587067Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-59c77dc479-mgw2d,Uid:2144b56e-33de-4010-bbdd-c43fd2475741,Namespace:calico-system,Attempt:0,} returns sandbox id \"eef28cfd2d2788309278cf3b1cee153fe7737cfef83c835b2f5d2436cd61b358\"" Mar 7 01:13:40.841534 containerd[1638]: time="2026-03-07T01:13:40.840966664Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\"" Mar 7 01:13:40.845371 kubelet[2747]: E0307 01:13:40.845035 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.845371 kubelet[2747]: W0307 01:13:40.845047 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.845371 kubelet[2747]: E0307 01:13:40.845056 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.930392 kubelet[2747]: E0307 01:13:40.930334 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.930392 kubelet[2747]: W0307 01:13:40.930360 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.930392 kubelet[2747]: E0307 01:13:40.930383 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.930945 kubelet[2747]: E0307 01:13:40.930916 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.930945 kubelet[2747]: W0307 01:13:40.930936 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.931012 kubelet[2747]: E0307 01:13:40.930950 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.931473 kubelet[2747]: E0307 01:13:40.931427 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.931473 kubelet[2747]: W0307 01:13:40.931455 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.931473 kubelet[2747]: E0307 01:13:40.931480 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.932171 kubelet[2747]: E0307 01:13:40.932136 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.932171 kubelet[2747]: W0307 01:13:40.932159 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.932375 kubelet[2747]: E0307 01:13:40.932173 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.932601 kubelet[2747]: E0307 01:13:40.932536 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.932601 kubelet[2747]: W0307 01:13:40.932598 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.932750 kubelet[2747]: E0307 01:13:40.932611 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.933027 kubelet[2747]: E0307 01:13:40.932955 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.933027 kubelet[2747]: W0307 01:13:40.932970 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.933027 kubelet[2747]: E0307 01:13:40.932982 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.933800 kubelet[2747]: E0307 01:13:40.933333 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.933800 kubelet[2747]: W0307 01:13:40.933384 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.933800 kubelet[2747]: E0307 01:13:40.933395 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.933965 kubelet[2747]: E0307 01:13:40.933914 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.933965 kubelet[2747]: W0307 01:13:40.933927 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.933965 kubelet[2747]: E0307 01:13:40.933954 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.934466 kubelet[2747]: E0307 01:13:40.934406 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.934466 kubelet[2747]: W0307 01:13:40.934423 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.934466 kubelet[2747]: E0307 01:13:40.934436 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.935344 kubelet[2747]: E0307 01:13:40.934814 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.935344 kubelet[2747]: W0307 01:13:40.934835 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.935344 kubelet[2747]: E0307 01:13:40.934851 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.935344 kubelet[2747]: E0307 01:13:40.935223 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.935344 kubelet[2747]: W0307 01:13:40.935241 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.935344 kubelet[2747]: E0307 01:13:40.935255 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.936149 kubelet[2747]: E0307 01:13:40.935922 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.936149 kubelet[2747]: W0307 01:13:40.935938 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.936149 kubelet[2747]: E0307 01:13:40.935951 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.936360 kubelet[2747]: E0307 01:13:40.936330 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.936454 kubelet[2747]: W0307 01:13:40.936348 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.936454 kubelet[2747]: E0307 01:13:40.936447 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.936913 kubelet[2747]: E0307 01:13:40.936870 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.936913 kubelet[2747]: W0307 01:13:40.936887 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.936913 kubelet[2747]: E0307 01:13:40.936915 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.937354 kubelet[2747]: E0307 01:13:40.937319 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.937354 kubelet[2747]: W0307 01:13:40.937342 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.937478 kubelet[2747]: E0307 01:13:40.937364 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.938001 kubelet[2747]: E0307 01:13:40.937968 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.938001 kubelet[2747]: W0307 01:13:40.937990 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.938079 kubelet[2747]: E0307 01:13:40.938004 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.938411 kubelet[2747]: E0307 01:13:40.938379 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.938411 kubelet[2747]: W0307 01:13:40.938397 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.938411 kubelet[2747]: E0307 01:13:40.938409 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.938806 kubelet[2747]: E0307 01:13:40.938771 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.938806 kubelet[2747]: W0307 01:13:40.938796 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.938806 kubelet[2747]: E0307 01:13:40.938808 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.939350 kubelet[2747]: E0307 01:13:40.939312 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.939350 kubelet[2747]: W0307 01:13:40.939335 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.939350 kubelet[2747]: E0307 01:13:40.939349 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.939867 kubelet[2747]: E0307 01:13:40.939845 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.939867 kubelet[2747]: W0307 01:13:40.939863 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.939982 kubelet[2747]: E0307 01:13:40.939875 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.940342 kubelet[2747]: E0307 01:13:40.940296 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.940342 kubelet[2747]: W0307 01:13:40.940317 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.940342 kubelet[2747]: E0307 01:13:40.940332 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.941952 kubelet[2747]: E0307 01:13:40.941928 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.941952 kubelet[2747]: W0307 01:13:40.941949 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.942094 kubelet[2747]: E0307 01:13:40.941965 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.942394 kubelet[2747]: E0307 01:13:40.942357 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.942394 kubelet[2747]: W0307 01:13:40.942379 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.942394 kubelet[2747]: E0307 01:13:40.942394 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.943385 kubelet[2747]: E0307 01:13:40.943333 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.943385 kubelet[2747]: W0307 01:13:40.943354 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.943385 kubelet[2747]: E0307 01:13:40.943368 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.944810 kubelet[2747]: E0307 01:13:40.944728 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.944810 kubelet[2747]: W0307 01:13:40.944746 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.944810 kubelet[2747]: E0307 01:13:40.944761 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:40.951576 kubelet[2747]: E0307 01:13:40.951531 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:40.951679 kubelet[2747]: W0307 01:13:40.951646 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:40.951679 kubelet[2747]: E0307 01:13:40.951669 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:41.097707 containerd[1638]: time="2026-03-07T01:13:41.097630087Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-lscfk,Uid:eb326673-d2f8-44d7-8c06-d4f44d6af381,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:41.118477 containerd[1638]: time="2026-03-07T01:13:41.118339812Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:13:41.118477 containerd[1638]: time="2026-03-07T01:13:41.118379032Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:13:41.118477 containerd[1638]: time="2026-03-07T01:13:41.118388892Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:41.119093 containerd[1638]: time="2026-03-07T01:13:41.118569012Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:41.144975 containerd[1638]: time="2026-03-07T01:13:41.144939460Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-lscfk,Uid:eb326673-d2f8-44d7-8c06-d4f44d6af381,Namespace:calico-system,Attempt:0,} returns sandbox id \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\"" Mar 7 01:13:41.620393 systemd[1]: run-containerd-runc-k8s.io-eef28cfd2d2788309278cf3b1cee153fe7737cfef83c835b2f5d2436cd61b358-runc.AKfgoL.mount: Deactivated successfully. Mar 7 01:13:42.016139 kubelet[2747]: E0307 01:13:42.015966 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:42.476931 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3652440714.mount: Deactivated successfully. Mar 7 01:13:42.829661 containerd[1638]: time="2026-03-07T01:13:42.829558798Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:42.830609 containerd[1638]: time="2026-03-07T01:13:42.830574946Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.4: active requests=0, bytes read=36107596" Mar 7 01:13:42.831466 containerd[1638]: time="2026-03-07T01:13:42.831424395Z" level=info msg="ImageCreate event name:\"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:42.833215 containerd[1638]: time="2026-03-07T01:13:42.833185354Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:42.833794 containerd[1638]: time="2026-03-07T01:13:42.833777253Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.4\" with image id \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\", size \"36107450\" in 1.992788389s" Mar 7 01:13:42.833919 containerd[1638]: time="2026-03-07T01:13:42.833841673Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\" returns image reference \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\"" Mar 7 01:13:42.835465 containerd[1638]: time="2026-03-07T01:13:42.835439881Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\"" Mar 7 01:13:42.847170 containerd[1638]: time="2026-03-07T01:13:42.847016438Z" level=info msg="CreateContainer within sandbox \"eef28cfd2d2788309278cf3b1cee153fe7737cfef83c835b2f5d2436cd61b358\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Mar 7 01:13:42.861428 containerd[1638]: time="2026-03-07T01:13:42.861392823Z" level=info msg="CreateContainer within sandbox \"eef28cfd2d2788309278cf3b1cee153fe7737cfef83c835b2f5d2436cd61b358\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"3df4350828ccbaca4bdb384bb5243e747f4c7ff4aed8bea0b62d6c7a2fc6d195\"" Mar 7 01:13:42.862759 containerd[1638]: time="2026-03-07T01:13:42.862036612Z" level=info msg="StartContainer for \"3df4350828ccbaca4bdb384bb5243e747f4c7ff4aed8bea0b62d6c7a2fc6d195\"" Mar 7 01:13:42.923768 containerd[1638]: time="2026-03-07T01:13:42.923731914Z" level=info msg="StartContainer for \"3df4350828ccbaca4bdb384bb5243e747f4c7ff4aed8bea0b62d6c7a2fc6d195\" returns successfully" Mar 7 01:13:43.089053 kubelet[2747]: I0307 01:13:43.088731 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-59c77dc479-mgw2d" podStartSLOduration=1.094486163 podStartE2EDuration="3.08871979s" podCreationTimestamp="2026-03-07 01:13:40 +0000 UTC" firstStartedPulling="2026-03-07 01:13:40.840513985 +0000 UTC m=+15.907425702" lastFinishedPulling="2026-03-07 01:13:42.834747602 +0000 UTC m=+17.901659329" observedRunningTime="2026-03-07 01:13:43.088163219 +0000 UTC m=+18.155074936" watchObservedRunningTime="2026-03-07 01:13:43.08871979 +0000 UTC m=+18.155631517" Mar 7 01:13:43.112297 kubelet[2747]: E0307 01:13:43.112272 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.112297 kubelet[2747]: W0307 01:13:43.112296 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.112392 kubelet[2747]: E0307 01:13:43.112311 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.112636 kubelet[2747]: E0307 01:13:43.112603 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.112636 kubelet[2747]: W0307 01:13:43.112624 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.112636 kubelet[2747]: E0307 01:13:43.112645 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.112928 kubelet[2747]: E0307 01:13:43.112910 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.112928 kubelet[2747]: W0307 01:13:43.112918 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.112928 kubelet[2747]: E0307 01:13:43.112927 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.113275 kubelet[2747]: E0307 01:13:43.113250 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.113275 kubelet[2747]: W0307 01:13:43.113268 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.113323 kubelet[2747]: E0307 01:13:43.113276 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.113566 kubelet[2747]: E0307 01:13:43.113526 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.113566 kubelet[2747]: W0307 01:13:43.113559 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.113612 kubelet[2747]: E0307 01:13:43.113570 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.113868 kubelet[2747]: E0307 01:13:43.113846 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.113868 kubelet[2747]: W0307 01:13:43.113860 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.113868 kubelet[2747]: E0307 01:13:43.113868 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.114093 kubelet[2747]: E0307 01:13:43.114078 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.114093 kubelet[2747]: W0307 01:13:43.114088 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.114141 kubelet[2747]: E0307 01:13:43.114094 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.114320 kubelet[2747]: E0307 01:13:43.114307 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.114320 kubelet[2747]: W0307 01:13:43.114316 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.114355 kubelet[2747]: E0307 01:13:43.114322 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.114557 kubelet[2747]: E0307 01:13:43.114525 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.114557 kubelet[2747]: W0307 01:13:43.114535 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.114606 kubelet[2747]: E0307 01:13:43.114563 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.114786 kubelet[2747]: E0307 01:13:43.114765 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.114786 kubelet[2747]: W0307 01:13:43.114780 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.114786 kubelet[2747]: E0307 01:13:43.114785 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.115007 kubelet[2747]: E0307 01:13:43.114988 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.115007 kubelet[2747]: W0307 01:13:43.114998 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.115007 kubelet[2747]: E0307 01:13:43.115004 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.115231 kubelet[2747]: E0307 01:13:43.115213 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.115231 kubelet[2747]: W0307 01:13:43.115225 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.115231 kubelet[2747]: E0307 01:13:43.115231 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.115453 kubelet[2747]: E0307 01:13:43.115441 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.115453 kubelet[2747]: W0307 01:13:43.115450 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.115493 kubelet[2747]: E0307 01:13:43.115456 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.115692 kubelet[2747]: E0307 01:13:43.115681 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.115692 kubelet[2747]: W0307 01:13:43.115690 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.115734 kubelet[2747]: E0307 01:13:43.115713 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.115952 kubelet[2747]: E0307 01:13:43.115936 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.115952 kubelet[2747]: W0307 01:13:43.115946 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.115952 kubelet[2747]: E0307 01:13:43.115952 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.151033 kubelet[2747]: E0307 01:13:43.150949 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.151033 kubelet[2747]: W0307 01:13:43.150988 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.151239 kubelet[2747]: E0307 01:13:43.151052 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.152728 kubelet[2747]: E0307 01:13:43.152650 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.152728 kubelet[2747]: W0307 01:13:43.152682 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.153008 kubelet[2747]: E0307 01:13:43.152838 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.153718 kubelet[2747]: E0307 01:13:43.153672 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.153718 kubelet[2747]: W0307 01:13:43.153704 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.153831 kubelet[2747]: E0307 01:13:43.153723 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.154254 kubelet[2747]: E0307 01:13:43.154223 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.154254 kubelet[2747]: W0307 01:13:43.154245 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.154341 kubelet[2747]: E0307 01:13:43.154261 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.154748 kubelet[2747]: E0307 01:13:43.154718 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.154748 kubelet[2747]: W0307 01:13:43.154740 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.154822 kubelet[2747]: E0307 01:13:43.154755 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.155212 kubelet[2747]: E0307 01:13:43.155182 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.155212 kubelet[2747]: W0307 01:13:43.155203 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.155307 kubelet[2747]: E0307 01:13:43.155216 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.156213 kubelet[2747]: E0307 01:13:43.155693 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.156213 kubelet[2747]: W0307 01:13:43.155714 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.156213 kubelet[2747]: E0307 01:13:43.155729 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.156352 kubelet[2747]: E0307 01:13:43.156315 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.156352 kubelet[2747]: W0307 01:13:43.156336 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.156432 kubelet[2747]: E0307 01:13:43.156358 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.158150 kubelet[2747]: E0307 01:13:43.158105 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.158150 kubelet[2747]: W0307 01:13:43.158125 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.158150 kubelet[2747]: E0307 01:13:43.158143 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.158725 kubelet[2747]: E0307 01:13:43.158657 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.158725 kubelet[2747]: W0307 01:13:43.158669 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.158725 kubelet[2747]: E0307 01:13:43.158682 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.159193 kubelet[2747]: E0307 01:13:43.159161 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.159193 kubelet[2747]: W0307 01:13:43.159181 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.159193 kubelet[2747]: E0307 01:13:43.159192 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.159756 kubelet[2747]: E0307 01:13:43.159729 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.159756 kubelet[2747]: W0307 01:13:43.159747 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.159756 kubelet[2747]: E0307 01:13:43.159761 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.160261 kubelet[2747]: E0307 01:13:43.160233 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.160261 kubelet[2747]: W0307 01:13:43.160256 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.160362 kubelet[2747]: E0307 01:13:43.160268 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.160834 kubelet[2747]: E0307 01:13:43.160796 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.160834 kubelet[2747]: W0307 01:13:43.160823 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.160834 kubelet[2747]: E0307 01:13:43.160843 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.161712 kubelet[2747]: E0307 01:13:43.161652 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.161712 kubelet[2747]: W0307 01:13:43.161675 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.161712 kubelet[2747]: E0307 01:13:43.161689 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.162267 kubelet[2747]: E0307 01:13:43.162239 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.162347 kubelet[2747]: W0307 01:13:43.162291 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.162347 kubelet[2747]: E0307 01:13:43.162305 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.163043 kubelet[2747]: E0307 01:13:43.163016 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.163043 kubelet[2747]: W0307 01:13:43.163036 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.163496 kubelet[2747]: E0307 01:13:43.163050 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:43.163620 kubelet[2747]: E0307 01:13:43.163500 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:43.163620 kubelet[2747]: W0307 01:13:43.163511 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:43.163620 kubelet[2747]: E0307 01:13:43.163523 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.015655 kubelet[2747]: E0307 01:13:44.015467 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:44.080425 kubelet[2747]: I0307 01:13:44.080400 2747 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:13:44.122476 kubelet[2747]: E0307 01:13:44.122449 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.122476 kubelet[2747]: W0307 01:13:44.122465 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.122476 kubelet[2747]: E0307 01:13:44.122481 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.122861 kubelet[2747]: E0307 01:13:44.122704 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.122861 kubelet[2747]: W0307 01:13:44.122711 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.122861 kubelet[2747]: E0307 01:13:44.122718 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.122915 kubelet[2747]: E0307 01:13:44.122895 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.122915 kubelet[2747]: W0307 01:13:44.122903 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.122946 kubelet[2747]: E0307 01:13:44.122921 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.123130 kubelet[2747]: E0307 01:13:44.123115 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.123130 kubelet[2747]: W0307 01:13:44.123126 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.123172 kubelet[2747]: E0307 01:13:44.123133 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.123336 kubelet[2747]: E0307 01:13:44.123319 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.123336 kubelet[2747]: W0307 01:13:44.123330 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.123336 kubelet[2747]: E0307 01:13:44.123338 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.123580 kubelet[2747]: E0307 01:13:44.123530 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.123580 kubelet[2747]: W0307 01:13:44.123557 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.123580 kubelet[2747]: E0307 01:13:44.123566 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.123768 kubelet[2747]: E0307 01:13:44.123751 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.123768 kubelet[2747]: W0307 01:13:44.123761 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.123768 kubelet[2747]: E0307 01:13:44.123768 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.123987 kubelet[2747]: E0307 01:13:44.123970 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.123987 kubelet[2747]: W0307 01:13:44.123983 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.124027 kubelet[2747]: E0307 01:13:44.123992 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.124226 kubelet[2747]: E0307 01:13:44.124211 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.124226 kubelet[2747]: W0307 01:13:44.124221 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.124260 kubelet[2747]: E0307 01:13:44.124229 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.124442 kubelet[2747]: E0307 01:13:44.124427 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.124442 kubelet[2747]: W0307 01:13:44.124437 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.124442 kubelet[2747]: E0307 01:13:44.124444 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.124663 kubelet[2747]: E0307 01:13:44.124649 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.124663 kubelet[2747]: W0307 01:13:44.124659 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.124710 kubelet[2747]: E0307 01:13:44.124665 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.124856 kubelet[2747]: E0307 01:13:44.124838 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.124884 kubelet[2747]: W0307 01:13:44.124867 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.124884 kubelet[2747]: E0307 01:13:44.124874 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.125099 kubelet[2747]: E0307 01:13:44.125086 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.125099 kubelet[2747]: W0307 01:13:44.125094 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.125145 kubelet[2747]: E0307 01:13:44.125101 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.125376 kubelet[2747]: E0307 01:13:44.125291 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.125376 kubelet[2747]: W0307 01:13:44.125299 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.125376 kubelet[2747]: E0307 01:13:44.125306 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.125509 kubelet[2747]: E0307 01:13:44.125494 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.125509 kubelet[2747]: W0307 01:13:44.125503 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.125509 kubelet[2747]: E0307 01:13:44.125509 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.165010 kubelet[2747]: E0307 01:13:44.164973 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.165010 kubelet[2747]: W0307 01:13:44.164997 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.165010 kubelet[2747]: E0307 01:13:44.165014 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.165298 kubelet[2747]: E0307 01:13:44.165279 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.165298 kubelet[2747]: W0307 01:13:44.165293 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.165393 kubelet[2747]: E0307 01:13:44.165304 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.165626 kubelet[2747]: E0307 01:13:44.165561 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.165626 kubelet[2747]: W0307 01:13:44.165570 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.165626 kubelet[2747]: E0307 01:13:44.165578 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.165833 kubelet[2747]: E0307 01:13:44.165813 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.165833 kubelet[2747]: W0307 01:13:44.165826 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.165893 kubelet[2747]: E0307 01:13:44.165836 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.166088 kubelet[2747]: E0307 01:13:44.166072 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.166088 kubelet[2747]: W0307 01:13:44.166084 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.166145 kubelet[2747]: E0307 01:13:44.166093 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.166355 kubelet[2747]: E0307 01:13:44.166339 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.166355 kubelet[2747]: W0307 01:13:44.166350 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.166400 kubelet[2747]: E0307 01:13:44.166358 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.166636 kubelet[2747]: E0307 01:13:44.166619 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.166636 kubelet[2747]: W0307 01:13:44.166630 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.166636 kubelet[2747]: E0307 01:13:44.166637 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.167139 kubelet[2747]: E0307 01:13:44.167123 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.167139 kubelet[2747]: W0307 01:13:44.167135 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.167200 kubelet[2747]: E0307 01:13:44.167145 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.167451 kubelet[2747]: E0307 01:13:44.167434 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.167451 kubelet[2747]: W0307 01:13:44.167445 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.167451 kubelet[2747]: E0307 01:13:44.167452 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.167759 kubelet[2747]: E0307 01:13:44.167743 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.167759 kubelet[2747]: W0307 01:13:44.167754 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.167972 kubelet[2747]: E0307 01:13:44.167760 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.168171 kubelet[2747]: E0307 01:13:44.168033 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.168171 kubelet[2747]: W0307 01:13:44.168043 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.168171 kubelet[2747]: E0307 01:13:44.168051 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.168316 kubelet[2747]: E0307 01:13:44.168298 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.168316 kubelet[2747]: W0307 01:13:44.168310 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.168371 kubelet[2747]: E0307 01:13:44.168316 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.168591 kubelet[2747]: E0307 01:13:44.168568 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.168591 kubelet[2747]: W0307 01:13:44.168577 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.168591 kubelet[2747]: E0307 01:13:44.168585 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.168934 kubelet[2747]: E0307 01:13:44.168916 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.168934 kubelet[2747]: W0307 01:13:44.168929 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.168992 kubelet[2747]: E0307 01:13:44.168937 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.169183 kubelet[2747]: E0307 01:13:44.169163 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.169183 kubelet[2747]: W0307 01:13:44.169175 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.169228 kubelet[2747]: E0307 01:13:44.169184 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.169419 kubelet[2747]: E0307 01:13:44.169401 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.169419 kubelet[2747]: W0307 01:13:44.169411 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.169460 kubelet[2747]: E0307 01:13:44.169419 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.169833 kubelet[2747]: E0307 01:13:44.169819 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.169833 kubelet[2747]: W0307 01:13:44.169829 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.169876 kubelet[2747]: E0307 01:13:44.169837 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.170070 kubelet[2747]: E0307 01:13:44.170055 2747 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:13:44.170070 kubelet[2747]: W0307 01:13:44.170066 2747 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:13:44.170110 kubelet[2747]: E0307 01:13:44.170074 2747 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:13:44.379887 containerd[1638]: time="2026-03-07T01:13:44.379677606Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:44.382123 containerd[1638]: time="2026-03-07T01:13:44.381674735Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4: active requests=0, bytes read=4630250" Mar 7 01:13:44.382832 containerd[1638]: time="2026-03-07T01:13:44.382771144Z" level=info msg="ImageCreate event name:\"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:44.384889 containerd[1638]: time="2026-03-07T01:13:44.384829242Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:44.385819 containerd[1638]: time="2026-03-07T01:13:44.385376472Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" with image id \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\", size \"6186255\" in 1.549905341s" Mar 7 01:13:44.385819 containerd[1638]: time="2026-03-07T01:13:44.385403882Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" returns image reference \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\"" Mar 7 01:13:44.390097 containerd[1638]: time="2026-03-07T01:13:44.390062778Z" level=info msg="CreateContainer within sandbox \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 7 01:13:44.409654 containerd[1638]: time="2026-03-07T01:13:44.409604599Z" level=info msg="CreateContainer within sandbox \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"c84e929e6e942bfc45c5a49491d60af6c07bf2cd73bec1abc929015892196aa6\"" Mar 7 01:13:44.411130 containerd[1638]: time="2026-03-07T01:13:44.410217519Z" level=info msg="StartContainer for \"c84e929e6e942bfc45c5a49491d60af6c07bf2cd73bec1abc929015892196aa6\"" Mar 7 01:13:44.436520 systemd[1]: run-containerd-runc-k8s.io-c84e929e6e942bfc45c5a49491d60af6c07bf2cd73bec1abc929015892196aa6-runc.RaBmWV.mount: Deactivated successfully. Mar 7 01:13:44.465242 containerd[1638]: time="2026-03-07T01:13:44.465206588Z" level=info msg="StartContainer for \"c84e929e6e942bfc45c5a49491d60af6c07bf2cd73bec1abc929015892196aa6\" returns successfully" Mar 7 01:13:44.613086 containerd[1638]: time="2026-03-07T01:13:44.613017691Z" level=info msg="shim disconnected" id=c84e929e6e942bfc45c5a49491d60af6c07bf2cd73bec1abc929015892196aa6 namespace=k8s.io Mar 7 01:13:44.613086 containerd[1638]: time="2026-03-07T01:13:44.613074951Z" level=warning msg="cleaning up after shim disconnected" id=c84e929e6e942bfc45c5a49491d60af6c07bf2cd73bec1abc929015892196aa6 namespace=k8s.io Mar 7 01:13:44.613086 containerd[1638]: time="2026-03-07T01:13:44.613082671Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:13:45.083383 containerd[1638]: time="2026-03-07T01:13:45.083335054Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\"" Mar 7 01:13:45.399985 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c84e929e6e942bfc45c5a49491d60af6c07bf2cd73bec1abc929015892196aa6-rootfs.mount: Deactivated successfully. Mar 7 01:13:46.015839 kubelet[2747]: E0307 01:13:46.015775 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:48.015801 kubelet[2747]: E0307 01:13:48.015735 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:49.036763 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2940082138.mount: Deactivated successfully. Mar 7 01:13:49.063036 containerd[1638]: time="2026-03-07T01:13:49.062962654Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:49.064302 containerd[1638]: time="2026-03-07T01:13:49.064262743Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.4: active requests=0, bytes read=159838564" Mar 7 01:13:49.065360 containerd[1638]: time="2026-03-07T01:13:49.065055423Z" level=info msg="ImageCreate event name:\"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:49.067236 containerd[1638]: time="2026-03-07T01:13:49.067208932Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:49.067708 containerd[1638]: time="2026-03-07T01:13:49.067683062Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.4\" with image id \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\", size \"159838426\" in 3.984315148s" Mar 7 01:13:49.067764 containerd[1638]: time="2026-03-07T01:13:49.067754262Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\" returns image reference \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\"" Mar 7 01:13:49.071080 containerd[1638]: time="2026-03-07T01:13:49.071063090Z" level=info msg="CreateContainer within sandbox \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Mar 7 01:13:49.098449 containerd[1638]: time="2026-03-07T01:13:49.098398935Z" level=info msg="CreateContainer within sandbox \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"27514627cd3350a7ca7163f7f545952653635f2a6a4145b8a6a2493f36a8018a\"" Mar 7 01:13:49.098971 containerd[1638]: time="2026-03-07T01:13:49.098826295Z" level=info msg="StartContainer for \"27514627cd3350a7ca7163f7f545952653635f2a6a4145b8a6a2493f36a8018a\"" Mar 7 01:13:49.154916 containerd[1638]: time="2026-03-07T01:13:49.154877563Z" level=info msg="StartContainer for \"27514627cd3350a7ca7163f7f545952653635f2a6a4145b8a6a2493f36a8018a\" returns successfully" Mar 7 01:13:49.264397 containerd[1638]: time="2026-03-07T01:13:49.264322143Z" level=info msg="shim disconnected" id=27514627cd3350a7ca7163f7f545952653635f2a6a4145b8a6a2493f36a8018a namespace=k8s.io Mar 7 01:13:49.264397 containerd[1638]: time="2026-03-07T01:13:49.264387883Z" level=warning msg="cleaning up after shim disconnected" id=27514627cd3350a7ca7163f7f545952653635f2a6a4145b8a6a2493f36a8018a namespace=k8s.io Mar 7 01:13:49.264397 containerd[1638]: time="2026-03-07T01:13:49.264395353Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:13:50.015459 kubelet[2747]: E0307 01:13:50.015373 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:50.037911 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-27514627cd3350a7ca7163f7f545952653635f2a6a4145b8a6a2493f36a8018a-rootfs.mount: Deactivated successfully. Mar 7 01:13:50.101398 containerd[1638]: time="2026-03-07T01:13:50.100169736Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\"" Mar 7 01:13:52.015571 kubelet[2747]: E0307 01:13:52.015240 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:52.668612 containerd[1638]: time="2026-03-07T01:13:52.668569920Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:52.669354 containerd[1638]: time="2026-03-07T01:13:52.669250219Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.4: active requests=0, bytes read=70611671" Mar 7 01:13:52.669979 containerd[1638]: time="2026-03-07T01:13:52.669937319Z" level=info msg="ImageCreate event name:\"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:52.672179 containerd[1638]: time="2026-03-07T01:13:52.671587609Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:52.672179 containerd[1638]: time="2026-03-07T01:13:52.672088268Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.4\" with image id \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\", size \"72167716\" in 2.571885112s" Mar 7 01:13:52.672179 containerd[1638]: time="2026-03-07T01:13:52.672115458Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\" returns image reference \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\"" Mar 7 01:13:52.675963 containerd[1638]: time="2026-03-07T01:13:52.675945887Z" level=info msg="CreateContainer within sandbox \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 7 01:13:52.688073 containerd[1638]: time="2026-03-07T01:13:52.688040102Z" level=info msg="CreateContainer within sandbox \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"100801284cd139388a4ab0efaea37d5138a38d7240112ebec62f22f29897f44d\"" Mar 7 01:13:52.688646 containerd[1638]: time="2026-03-07T01:13:52.688566782Z" level=info msg="StartContainer for \"100801284cd139388a4ab0efaea37d5138a38d7240112ebec62f22f29897f44d\"" Mar 7 01:13:52.711456 systemd[1]: run-containerd-runc-k8s.io-100801284cd139388a4ab0efaea37d5138a38d7240112ebec62f22f29897f44d-runc.a1E7yx.mount: Deactivated successfully. Mar 7 01:13:52.738607 containerd[1638]: time="2026-03-07T01:13:52.738534693Z" level=info msg="StartContainer for \"100801284cd139388a4ab0efaea37d5138a38d7240112ebec62f22f29897f44d\" returns successfully" Mar 7 01:13:53.214002 containerd[1638]: time="2026-03-07T01:13:53.213832832Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 7 01:13:53.237987 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-100801284cd139388a4ab0efaea37d5138a38d7240112ebec62f22f29897f44d-rootfs.mount: Deactivated successfully. Mar 7 01:13:53.239239 containerd[1638]: time="2026-03-07T01:13:53.239016453Z" level=info msg="shim disconnected" id=100801284cd139388a4ab0efaea37d5138a38d7240112ebec62f22f29897f44d namespace=k8s.io Mar 7 01:13:53.239239 containerd[1638]: time="2026-03-07T01:13:53.239068813Z" level=warning msg="cleaning up after shim disconnected" id=100801284cd139388a4ab0efaea37d5138a38d7240112ebec62f22f29897f44d namespace=k8s.io Mar 7 01:13:53.239239 containerd[1638]: time="2026-03-07T01:13:53.239077183Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:13:53.272723 kubelet[2747]: I0307 01:13:53.272676 2747 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Mar 7 01:13:53.426762 kubelet[2747]: I0307 01:13:53.426694 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/70b07735-6f2b-40bc-8301-190b364be2aa-whisker-backend-key-pair\") pod \"whisker-687768496d-8mqt2\" (UID: \"70b07735-6f2b-40bc-8301-190b364be2aa\") " pod="calico-system/whisker-687768496d-8mqt2" Mar 7 01:13:53.426762 kubelet[2747]: I0307 01:13:53.426742 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70b07735-6f2b-40bc-8301-190b364be2aa-whisker-ca-bundle\") pod \"whisker-687768496d-8mqt2\" (UID: \"70b07735-6f2b-40bc-8301-190b364be2aa\") " pod="calico-system/whisker-687768496d-8mqt2" Mar 7 01:13:53.426762 kubelet[2747]: I0307 01:13:53.426759 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngk9n\" (UniqueName: \"kubernetes.io/projected/70b07735-6f2b-40bc-8301-190b364be2aa-kube-api-access-ngk9n\") pod \"whisker-687768496d-8mqt2\" (UID: \"70b07735-6f2b-40bc-8301-190b364be2aa\") " pod="calico-system/whisker-687768496d-8mqt2" Mar 7 01:13:53.426762 kubelet[2747]: I0307 01:13:53.426776 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5g9s\" (UniqueName: \"kubernetes.io/projected/7f399497-ed30-4f43-b4f2-62919a08223a-kube-api-access-x5g9s\") pod \"calico-kube-controllers-5dc64846bc-k762g\" (UID: \"7f399497-ed30-4f43-b4f2-62919a08223a\") " pod="calico-system/calico-kube-controllers-5dc64846bc-k762g" Mar 7 01:13:53.426998 kubelet[2747]: I0307 01:13:53.426795 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pldzm\" (UniqueName: \"kubernetes.io/projected/99f62e43-c05f-4b23-90d0-600585809c83-kube-api-access-pldzm\") pod \"coredns-674b8bbfcf-4xvx5\" (UID: \"99f62e43-c05f-4b23-90d0-600585809c83\") " pod="kube-system/coredns-674b8bbfcf-4xvx5" Mar 7 01:13:53.426998 kubelet[2747]: I0307 01:13:53.426810 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/571d5c28-cf64-4799-a2c5-6c5bf2b8ab06-calico-apiserver-certs\") pod \"calico-apiserver-689889c887-zgnj6\" (UID: \"571d5c28-cf64-4799-a2c5-6c5bf2b8ab06\") " pod="calico-system/calico-apiserver-689889c887-zgnj6" Mar 7 01:13:53.426998 kubelet[2747]: I0307 01:13:53.426828 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v7p8\" (UniqueName: \"kubernetes.io/projected/a9f53e47-46dd-420a-9df1-28f9cae95a81-kube-api-access-9v7p8\") pod \"calico-apiserver-689889c887-ntpwf\" (UID: \"a9f53e47-46dd-420a-9df1-28f9cae95a81\") " pod="calico-system/calico-apiserver-689889c887-ntpwf" Mar 7 01:13:53.426998 kubelet[2747]: I0307 01:13:53.426844 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/70b07735-6f2b-40bc-8301-190b364be2aa-nginx-config\") pod \"whisker-687768496d-8mqt2\" (UID: \"70b07735-6f2b-40bc-8301-190b364be2aa\") " pod="calico-system/whisker-687768496d-8mqt2" Mar 7 01:13:53.426998 kubelet[2747]: I0307 01:13:53.426865 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gssvx\" (UniqueName: \"kubernetes.io/projected/1d72866e-1f81-4c34-a51f-96e1e7fc603a-kube-api-access-gssvx\") pod \"coredns-674b8bbfcf-98kpm\" (UID: \"1d72866e-1f81-4c34-a51f-96e1e7fc603a\") " pod="kube-system/coredns-674b8bbfcf-98kpm" Mar 7 01:13:53.427089 kubelet[2747]: I0307 01:13:53.426882 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99f62e43-c05f-4b23-90d0-600585809c83-config-volume\") pod \"coredns-674b8bbfcf-4xvx5\" (UID: \"99f62e43-c05f-4b23-90d0-600585809c83\") " pod="kube-system/coredns-674b8bbfcf-4xvx5" Mar 7 01:13:53.427089 kubelet[2747]: I0307 01:13:53.426900 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/a9f53e47-46dd-420a-9df1-28f9cae95a81-calico-apiserver-certs\") pod \"calico-apiserver-689889c887-ntpwf\" (UID: \"a9f53e47-46dd-420a-9df1-28f9cae95a81\") " pod="calico-system/calico-apiserver-689889c887-ntpwf" Mar 7 01:13:53.427089 kubelet[2747]: I0307 01:13:53.426919 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca853831-8f8c-4ad5-a296-b442696cf9d1-config\") pod \"goldmane-5b85766d88-9qlr7\" (UID: \"ca853831-8f8c-4ad5-a296-b442696cf9d1\") " pod="calico-system/goldmane-5b85766d88-9qlr7" Mar 7 01:13:53.427089 kubelet[2747]: I0307 01:13:53.426947 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca853831-8f8c-4ad5-a296-b442696cf9d1-goldmane-ca-bundle\") pod \"goldmane-5b85766d88-9qlr7\" (UID: \"ca853831-8f8c-4ad5-a296-b442696cf9d1\") " pod="calico-system/goldmane-5b85766d88-9qlr7" Mar 7 01:13:53.427089 kubelet[2747]: I0307 01:13:53.426962 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/ca853831-8f8c-4ad5-a296-b442696cf9d1-goldmane-key-pair\") pod \"goldmane-5b85766d88-9qlr7\" (UID: \"ca853831-8f8c-4ad5-a296-b442696cf9d1\") " pod="calico-system/goldmane-5b85766d88-9qlr7" Mar 7 01:13:53.427180 kubelet[2747]: I0307 01:13:53.426977 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvq7b\" (UniqueName: \"kubernetes.io/projected/ca853831-8f8c-4ad5-a296-b442696cf9d1-kube-api-access-tvq7b\") pod \"goldmane-5b85766d88-9qlr7\" (UID: \"ca853831-8f8c-4ad5-a296-b442696cf9d1\") " pod="calico-system/goldmane-5b85766d88-9qlr7" Mar 7 01:13:53.427180 kubelet[2747]: I0307 01:13:53.426995 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d72866e-1f81-4c34-a51f-96e1e7fc603a-config-volume\") pod \"coredns-674b8bbfcf-98kpm\" (UID: \"1d72866e-1f81-4c34-a51f-96e1e7fc603a\") " pod="kube-system/coredns-674b8bbfcf-98kpm" Mar 7 01:13:53.427180 kubelet[2747]: I0307 01:13:53.427010 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncz7l\" (UniqueName: \"kubernetes.io/projected/571d5c28-cf64-4799-a2c5-6c5bf2b8ab06-kube-api-access-ncz7l\") pod \"calico-apiserver-689889c887-zgnj6\" (UID: \"571d5c28-cf64-4799-a2c5-6c5bf2b8ab06\") " pod="calico-system/calico-apiserver-689889c887-zgnj6" Mar 7 01:13:53.427180 kubelet[2747]: I0307 01:13:53.427029 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f399497-ed30-4f43-b4f2-62919a08223a-tigera-ca-bundle\") pod \"calico-kube-controllers-5dc64846bc-k762g\" (UID: \"7f399497-ed30-4f43-b4f2-62919a08223a\") " pod="calico-system/calico-kube-controllers-5dc64846bc-k762g" Mar 7 01:13:53.613794 containerd[1638]: time="2026-03-07T01:13:53.613604128Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-687768496d-8mqt2,Uid:70b07735-6f2b-40bc-8301-190b364be2aa,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:53.619900 containerd[1638]: time="2026-03-07T01:13:53.619739707Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5dc64846bc-k762g,Uid:7f399497-ed30-4f43-b4f2-62919a08223a,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:53.623750 containerd[1638]: time="2026-03-07T01:13:53.623697615Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-98kpm,Uid:1d72866e-1f81-4c34-a51f-96e1e7fc603a,Namespace:kube-system,Attempt:0,}" Mar 7 01:13:53.626372 containerd[1638]: time="2026-03-07T01:13:53.626315684Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-9qlr7,Uid:ca853831-8f8c-4ad5-a296-b442696cf9d1,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:53.626754 containerd[1638]: time="2026-03-07T01:13:53.626702015Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-4xvx5,Uid:99f62e43-c05f-4b23-90d0-600585809c83,Namespace:kube-system,Attempt:0,}" Mar 7 01:13:53.632277 containerd[1638]: time="2026-03-07T01:13:53.632193782Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-689889c887-zgnj6,Uid:571d5c28-cf64-4799-a2c5-6c5bf2b8ab06,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:53.633895 containerd[1638]: time="2026-03-07T01:13:53.633805422Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-689889c887-ntpwf,Uid:a9f53e47-46dd-420a-9df1-28f9cae95a81,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:53.822691 containerd[1638]: time="2026-03-07T01:13:53.821745029Z" level=error msg="Failed to destroy network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.824478 containerd[1638]: time="2026-03-07T01:13:53.824449918Z" level=error msg="encountered an error cleaning up failed sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.824521 containerd[1638]: time="2026-03-07T01:13:53.824491938Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-9qlr7,Uid:ca853831-8f8c-4ad5-a296-b442696cf9d1,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.825004 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc-shm.mount: Deactivated successfully. Mar 7 01:13:53.825574 kubelet[2747]: E0307 01:13:53.825526 2747 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.825672 kubelet[2747]: E0307 01:13:53.825601 2747 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-5b85766d88-9qlr7" Mar 7 01:13:53.825672 kubelet[2747]: E0307 01:13:53.825619 2747 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-5b85766d88-9qlr7" Mar 7 01:13:53.825816 kubelet[2747]: E0307 01:13:53.825660 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-5b85766d88-9qlr7_calico-system(ca853831-8f8c-4ad5-a296-b442696cf9d1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-5b85766d88-9qlr7_calico-system(ca853831-8f8c-4ad5-a296-b442696cf9d1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-5b85766d88-9qlr7" podUID="ca853831-8f8c-4ad5-a296-b442696cf9d1" Mar 7 01:13:53.835667 containerd[1638]: time="2026-03-07T01:13:53.835627485Z" level=error msg="Failed to destroy network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.836651 containerd[1638]: time="2026-03-07T01:13:53.836059725Z" level=error msg="encountered an error cleaning up failed sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.836651 containerd[1638]: time="2026-03-07T01:13:53.836108665Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-4xvx5,Uid:99f62e43-c05f-4b23-90d0-600585809c83,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.838554 kubelet[2747]: E0307 01:13:53.837624 2747 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.838746 kubelet[2747]: E0307 01:13:53.838640 2747 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-4xvx5" Mar 7 01:13:53.838746 kubelet[2747]: E0307 01:13:53.838662 2747 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-4xvx5" Mar 7 01:13:53.838746 kubelet[2747]: E0307 01:13:53.838708 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-4xvx5_kube-system(99f62e43-c05f-4b23-90d0-600585809c83)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-4xvx5_kube-system(99f62e43-c05f-4b23-90d0-600585809c83)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-4xvx5" podUID="99f62e43-c05f-4b23-90d0-600585809c83" Mar 7 01:13:53.840088 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339-shm.mount: Deactivated successfully. Mar 7 01:13:53.851309 containerd[1638]: time="2026-03-07T01:13:53.851191439Z" level=error msg="Failed to destroy network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.852691 containerd[1638]: time="2026-03-07T01:13:53.851643119Z" level=error msg="encountered an error cleaning up failed sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.852691 containerd[1638]: time="2026-03-07T01:13:53.852607789Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-687768496d-8mqt2,Uid:70b07735-6f2b-40bc-8301-190b364be2aa,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.854605 kubelet[2747]: E0307 01:13:53.853669 2747 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.854605 kubelet[2747]: E0307 01:13:53.853724 2747 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-687768496d-8mqt2" Mar 7 01:13:53.854605 kubelet[2747]: E0307 01:13:53.853741 2747 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-687768496d-8mqt2" Mar 7 01:13:53.854691 kubelet[2747]: E0307 01:13:53.853778 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-687768496d-8mqt2_calico-system(70b07735-6f2b-40bc-8301-190b364be2aa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-687768496d-8mqt2_calico-system(70b07735-6f2b-40bc-8301-190b364be2aa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-687768496d-8mqt2" podUID="70b07735-6f2b-40bc-8301-190b364be2aa" Mar 7 01:13:53.856165 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b-shm.mount: Deactivated successfully. Mar 7 01:13:53.857780 containerd[1638]: time="2026-03-07T01:13:53.857691447Z" level=error msg="Failed to destroy network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.858107 containerd[1638]: time="2026-03-07T01:13:53.858068397Z" level=error msg="encountered an error cleaning up failed sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.858224 containerd[1638]: time="2026-03-07T01:13:53.858166697Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5dc64846bc-k762g,Uid:7f399497-ed30-4f43-b4f2-62919a08223a,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.858471 kubelet[2747]: E0307 01:13:53.858450 2747 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.859069 kubelet[2747]: E0307 01:13:53.858672 2747 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5dc64846bc-k762g" Mar 7 01:13:53.859069 kubelet[2747]: E0307 01:13:53.858694 2747 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5dc64846bc-k762g" Mar 7 01:13:53.859069 kubelet[2747]: E0307 01:13:53.858749 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5dc64846bc-k762g_calico-system(7f399497-ed30-4f43-b4f2-62919a08223a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5dc64846bc-k762g_calico-system(7f399497-ed30-4f43-b4f2-62919a08223a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5dc64846bc-k762g" podUID="7f399497-ed30-4f43-b4f2-62919a08223a" Mar 7 01:13:53.865579 containerd[1638]: time="2026-03-07T01:13:53.865092255Z" level=error msg="Failed to destroy network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.866484 containerd[1638]: time="2026-03-07T01:13:53.866462254Z" level=error msg="encountered an error cleaning up failed sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.866610 containerd[1638]: time="2026-03-07T01:13:53.866595254Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-98kpm,Uid:1d72866e-1f81-4c34-a51f-96e1e7fc603a,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.867093 kubelet[2747]: E0307 01:13:53.866813 2747 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.867093 kubelet[2747]: E0307 01:13:53.866857 2747 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-98kpm" Mar 7 01:13:53.867093 kubelet[2747]: E0307 01:13:53.866872 2747 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-98kpm" Mar 7 01:13:53.867191 kubelet[2747]: E0307 01:13:53.866911 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-98kpm_kube-system(1d72866e-1f81-4c34-a51f-96e1e7fc603a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-98kpm_kube-system(1d72866e-1f81-4c34-a51f-96e1e7fc603a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-98kpm" podUID="1d72866e-1f81-4c34-a51f-96e1e7fc603a" Mar 7 01:13:53.878281 containerd[1638]: time="2026-03-07T01:13:53.878243450Z" level=error msg="Failed to destroy network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.878790 containerd[1638]: time="2026-03-07T01:13:53.878649690Z" level=error msg="encountered an error cleaning up failed sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.878790 containerd[1638]: time="2026-03-07T01:13:53.878691131Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-689889c887-ntpwf,Uid:a9f53e47-46dd-420a-9df1-28f9cae95a81,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.879217 kubelet[2747]: E0307 01:13:53.878836 2747 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.879217 kubelet[2747]: E0307 01:13:53.878873 2747 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-689889c887-ntpwf" Mar 7 01:13:53.879217 kubelet[2747]: E0307 01:13:53.878899 2747 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-689889c887-ntpwf" Mar 7 01:13:53.879289 kubelet[2747]: E0307 01:13:53.878949 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-689889c887-ntpwf_calico-system(a9f53e47-46dd-420a-9df1-28f9cae95a81)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-689889c887-ntpwf_calico-system(a9f53e47-46dd-420a-9df1-28f9cae95a81)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-689889c887-ntpwf" podUID="a9f53e47-46dd-420a-9df1-28f9cae95a81" Mar 7 01:13:53.880292 containerd[1638]: time="2026-03-07T01:13:53.879953900Z" level=error msg="Failed to destroy network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.880292 containerd[1638]: time="2026-03-07T01:13:53.880204140Z" level=error msg="encountered an error cleaning up failed sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.880292 containerd[1638]: time="2026-03-07T01:13:53.880238160Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-689889c887-zgnj6,Uid:571d5c28-cf64-4799-a2c5-6c5bf2b8ab06,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.880378 kubelet[2747]: E0307 01:13:53.880346 2747 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:53.880378 kubelet[2747]: E0307 01:13:53.880368 2747 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-689889c887-zgnj6" Mar 7 01:13:53.880421 kubelet[2747]: E0307 01:13:53.880380 2747 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-689889c887-zgnj6" Mar 7 01:13:53.880421 kubelet[2747]: E0307 01:13:53.880404 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-689889c887-zgnj6_calico-system(571d5c28-cf64-4799-a2c5-6c5bf2b8ab06)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-689889c887-zgnj6_calico-system(571d5c28-cf64-4799-a2c5-6c5bf2b8ab06)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-689889c887-zgnj6" podUID="571d5c28-cf64-4799-a2c5-6c5bf2b8ab06" Mar 7 01:13:54.021208 containerd[1638]: time="2026-03-07T01:13:54.020711354Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-xbz8b,Uid:1fd3aff8-915e-4a01-9844-abd28ee38035,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:54.105808 containerd[1638]: time="2026-03-07T01:13:54.105727280Z" level=error msg="Failed to destroy network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.106573 containerd[1638]: time="2026-03-07T01:13:54.106504359Z" level=error msg="encountered an error cleaning up failed sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.106709 containerd[1638]: time="2026-03-07T01:13:54.106670199Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-xbz8b,Uid:1fd3aff8-915e-4a01-9844-abd28ee38035,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.106966 kubelet[2747]: E0307 01:13:54.106934 2747 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.107075 kubelet[2747]: E0307 01:13:54.107037 2747 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-xbz8b" Mar 7 01:13:54.107075 kubelet[2747]: E0307 01:13:54.107060 2747 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-xbz8b" Mar 7 01:13:54.107221 kubelet[2747]: E0307 01:13:54.107105 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-xbz8b_calico-system(1fd3aff8-915e-4a01-9844-abd28ee38035)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-xbz8b_calico-system(1fd3aff8-915e-4a01-9844-abd28ee38035)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:54.110810 kubelet[2747]: I0307 01:13:54.110785 2747 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:13:54.111393 containerd[1638]: time="2026-03-07T01:13:54.111364238Z" level=info msg="StopPodSandbox for \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\"" Mar 7 01:13:54.112003 containerd[1638]: time="2026-03-07T01:13:54.111480268Z" level=info msg="Ensure that sandbox 3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb in task-service has been cleanup successfully" Mar 7 01:13:54.114152 kubelet[2747]: I0307 01:13:54.114038 2747 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:13:54.114688 containerd[1638]: time="2026-03-07T01:13:54.114398008Z" level=info msg="StopPodSandbox for \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\"" Mar 7 01:13:54.114688 containerd[1638]: time="2026-03-07T01:13:54.114513887Z" level=info msg="Ensure that sandbox dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e in task-service has been cleanup successfully" Mar 7 01:13:54.116387 kubelet[2747]: I0307 01:13:54.116183 2747 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:13:54.116843 containerd[1638]: time="2026-03-07T01:13:54.116679487Z" level=info msg="StopPodSandbox for \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\"" Mar 7 01:13:54.119463 containerd[1638]: time="2026-03-07T01:13:54.117110957Z" level=info msg="Ensure that sandbox af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b in task-service has been cleanup successfully" Mar 7 01:13:54.131580 kubelet[2747]: I0307 01:13:54.131248 2747 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:13:54.139851 containerd[1638]: time="2026-03-07T01:13:54.138084891Z" level=info msg="StopPodSandbox for \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\"" Mar 7 01:13:54.140029 containerd[1638]: time="2026-03-07T01:13:54.140012830Z" level=info msg="Ensure that sandbox e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd in task-service has been cleanup successfully" Mar 7 01:13:54.147381 containerd[1638]: time="2026-03-07T01:13:54.146977258Z" level=info msg="CreateContainer within sandbox \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 7 01:13:54.148487 kubelet[2747]: I0307 01:13:54.148473 2747 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:13:54.151039 containerd[1638]: time="2026-03-07T01:13:54.151019697Z" level=info msg="StopPodSandbox for \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\"" Mar 7 01:13:54.151192 kubelet[2747]: I0307 01:13:54.151180 2747 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:13:54.152183 kubelet[2747]: I0307 01:13:54.152172 2747 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:13:54.153847 containerd[1638]: time="2026-03-07T01:13:54.153353206Z" level=info msg="StopPodSandbox for \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\"" Mar 7 01:13:54.154039 containerd[1638]: time="2026-03-07T01:13:54.153974806Z" level=info msg="Ensure that sandbox 37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971 in task-service has been cleanup successfully" Mar 7 01:13:54.154328 containerd[1638]: time="2026-03-07T01:13:54.154313736Z" level=info msg="Ensure that sandbox 6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1 in task-service has been cleanup successfully" Mar 7 01:13:54.157583 containerd[1638]: time="2026-03-07T01:13:54.151888426Z" level=info msg="StopPodSandbox for \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\"" Mar 7 01:13:54.158066 containerd[1638]: time="2026-03-07T01:13:54.157669165Z" level=info msg="Ensure that sandbox e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339 in task-service has been cleanup successfully" Mar 7 01:13:54.170571 kubelet[2747]: I0307 01:13:54.170524 2747 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:13:54.172490 containerd[1638]: time="2026-03-07T01:13:54.172467611Z" level=info msg="StopPodSandbox for \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\"" Mar 7 01:13:54.172690 containerd[1638]: time="2026-03-07T01:13:54.172679411Z" level=info msg="Ensure that sandbox ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc in task-service has been cleanup successfully" Mar 7 01:13:54.194791 containerd[1638]: time="2026-03-07T01:13:54.194751364Z" level=info msg="CreateContainer within sandbox \"08d623fa0848169e1eabab1deb2f726011d057a4f02d6b631bc72be04d0364f1\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"4d6f2b013a064a1d00774969e13f724b0eda1d044831eed38b7c44ff377ebbc2\"" Mar 7 01:13:54.196416 containerd[1638]: time="2026-03-07T01:13:54.195991614Z" level=info msg="StartContainer for \"4d6f2b013a064a1d00774969e13f724b0eda1d044831eed38b7c44ff377ebbc2\"" Mar 7 01:13:54.210307 containerd[1638]: time="2026-03-07T01:13:54.210271560Z" level=error msg="StopPodSandbox for \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\" failed" error="failed to destroy network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.210750 kubelet[2747]: E0307 01:13:54.210619 2747 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:13:54.210750 kubelet[2747]: E0307 01:13:54.210666 2747 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b"} Mar 7 01:13:54.210750 kubelet[2747]: E0307 01:13:54.210706 2747 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"70b07735-6f2b-40bc-8301-190b364be2aa\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:13:54.210750 kubelet[2747]: E0307 01:13:54.210726 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"70b07735-6f2b-40bc-8301-190b364be2aa\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-687768496d-8mqt2" podUID="70b07735-6f2b-40bc-8301-190b364be2aa" Mar 7 01:13:54.211142 containerd[1638]: time="2026-03-07T01:13:54.211122680Z" level=error msg="StopPodSandbox for \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\" failed" error="failed to destroy network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.211378 kubelet[2747]: E0307 01:13:54.211316 2747 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:13:54.211378 kubelet[2747]: E0307 01:13:54.211334 2747 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb"} Mar 7 01:13:54.211378 kubelet[2747]: E0307 01:13:54.211348 2747 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"1d72866e-1f81-4c34-a51f-96e1e7fc603a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:13:54.211378 kubelet[2747]: E0307 01:13:54.211361 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"1d72866e-1f81-4c34-a51f-96e1e7fc603a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-98kpm" podUID="1d72866e-1f81-4c34-a51f-96e1e7fc603a" Mar 7 01:13:54.218582 containerd[1638]: time="2026-03-07T01:13:54.218264948Z" level=error msg="StopPodSandbox for \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\" failed" error="failed to destroy network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.218879 kubelet[2747]: E0307 01:13:54.218749 2747 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:13:54.218879 kubelet[2747]: E0307 01:13:54.218823 2747 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e"} Mar 7 01:13:54.218879 kubelet[2747]: E0307 01:13:54.218846 2747 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"7f399497-ed30-4f43-b4f2-62919a08223a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:13:54.218879 kubelet[2747]: E0307 01:13:54.218860 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"7f399497-ed30-4f43-b4f2-62919a08223a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5dc64846bc-k762g" podUID="7f399497-ed30-4f43-b4f2-62919a08223a" Mar 7 01:13:54.243063 containerd[1638]: time="2026-03-07T01:13:54.243021770Z" level=error msg="StopPodSandbox for \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\" failed" error="failed to destroy network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.243560 kubelet[2747]: E0307 01:13:54.243424 2747 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:13:54.243560 kubelet[2747]: E0307 01:13:54.243461 2747 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971"} Mar 7 01:13:54.243560 kubelet[2747]: E0307 01:13:54.243492 2747 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"571d5c28-cf64-4799-a2c5-6c5bf2b8ab06\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:13:54.243560 kubelet[2747]: E0307 01:13:54.243511 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"571d5c28-cf64-4799-a2c5-6c5bf2b8ab06\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-689889c887-zgnj6" podUID="571d5c28-cf64-4799-a2c5-6c5bf2b8ab06" Mar 7 01:13:54.246485 containerd[1638]: time="2026-03-07T01:13:54.246450080Z" level=error msg="StopPodSandbox for \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\" failed" error="failed to destroy network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.246731 kubelet[2747]: E0307 01:13:54.246636 2747 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:13:54.246765 containerd[1638]: time="2026-03-07T01:13:54.246729060Z" level=error msg="StopPodSandbox for \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\" failed" error="failed to destroy network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.246786 kubelet[2747]: E0307 01:13:54.246665 2747 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd"} Mar 7 01:13:54.246805 kubelet[2747]: E0307 01:13:54.246786 2747 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"a9f53e47-46dd-420a-9df1-28f9cae95a81\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:13:54.246847 kubelet[2747]: E0307 01:13:54.246805 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"a9f53e47-46dd-420a-9df1-28f9cae95a81\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-689889c887-ntpwf" podUID="a9f53e47-46dd-420a-9df1-28f9cae95a81" Mar 7 01:13:54.247331 kubelet[2747]: E0307 01:13:54.247024 2747 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:13:54.247331 kubelet[2747]: E0307 01:13:54.247047 2747 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc"} Mar 7 01:13:54.247331 kubelet[2747]: E0307 01:13:54.247069 2747 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ca853831-8f8c-4ad5-a296-b442696cf9d1\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:13:54.247331 kubelet[2747]: E0307 01:13:54.247084 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ca853831-8f8c-4ad5-a296-b442696cf9d1\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-5b85766d88-9qlr7" podUID="ca853831-8f8c-4ad5-a296-b442696cf9d1" Mar 7 01:13:54.249246 containerd[1638]: time="2026-03-07T01:13:54.249226439Z" level=error msg="StopPodSandbox for \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\" failed" error="failed to destroy network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.249421 kubelet[2747]: E0307 01:13:54.249406 2747 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:13:54.249470 kubelet[2747]: E0307 01:13:54.249461 2747 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339"} Mar 7 01:13:54.249530 kubelet[2747]: E0307 01:13:54.249521 2747 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"99f62e43-c05f-4b23-90d0-600585809c83\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:13:54.249818 kubelet[2747]: E0307 01:13:54.249730 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"99f62e43-c05f-4b23-90d0-600585809c83\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-4xvx5" podUID="99f62e43-c05f-4b23-90d0-600585809c83" Mar 7 01:13:54.255531 containerd[1638]: time="2026-03-07T01:13:54.255501257Z" level=error msg="StopPodSandbox for \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\" failed" error="failed to destroy network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:13:54.255659 kubelet[2747]: E0307 01:13:54.255633 2747 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:13:54.255684 kubelet[2747]: E0307 01:13:54.255672 2747 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1"} Mar 7 01:13:54.255704 kubelet[2747]: E0307 01:13:54.255694 2747 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"1fd3aff8-915e-4a01-9844-abd28ee38035\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:13:54.255748 kubelet[2747]: E0307 01:13:54.255711 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"1fd3aff8-915e-4a01-9844-abd28ee38035\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-xbz8b" podUID="1fd3aff8-915e-4a01-9844-abd28ee38035" Mar 7 01:13:54.281684 containerd[1638]: time="2026-03-07T01:13:54.281653349Z" level=info msg="StartContainer for \"4d6f2b013a064a1d00774969e13f724b0eda1d044831eed38b7c44ff377ebbc2\" returns successfully" Mar 7 01:13:54.691918 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd-shm.mount: Deactivated successfully. Mar 7 01:13:54.692233 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971-shm.mount: Deactivated successfully. Mar 7 01:13:54.692474 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb-shm.mount: Deactivated successfully. Mar 7 01:13:54.693179 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e-shm.mount: Deactivated successfully. Mar 7 01:13:55.183419 containerd[1638]: time="2026-03-07T01:13:55.183374520Z" level=info msg="StopPodSandbox for \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\"" Mar 7 01:13:55.245655 kubelet[2747]: I0307 01:13:55.244576 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-lscfk" podStartSLOduration=3.717753907 podStartE2EDuration="15.244562355s" podCreationTimestamp="2026-03-07 01:13:40 +0000 UTC" firstStartedPulling="2026-03-07 01:13:41.146166159 +0000 UTC m=+16.213077876" lastFinishedPulling="2026-03-07 01:13:52.672974597 +0000 UTC m=+27.739886324" observedRunningTime="2026-03-07 01:13:55.210419633 +0000 UTC m=+30.277331380" watchObservedRunningTime="2026-03-07 01:13:55.244562355 +0000 UTC m=+30.311474072" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.243 [INFO][4090] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.244 [INFO][4090] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" iface="eth0" netns="/var/run/netns/cni-e8440e97-c3a8-3938-bb78-090bce3360c5" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.244 [INFO][4090] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" iface="eth0" netns="/var/run/netns/cni-e8440e97-c3a8-3938-bb78-090bce3360c5" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.244 [INFO][4090] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" iface="eth0" netns="/var/run/netns/cni-e8440e97-c3a8-3938-bb78-090bce3360c5" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.244 [INFO][4090] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.244 [INFO][4090] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.259 [INFO][4097] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.259 [INFO][4097] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.259 [INFO][4097] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.263 [WARNING][4097] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.263 [INFO][4097] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.264 [INFO][4097] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:13:55.269878 containerd[1638]: 2026-03-07 01:13:55.267 [INFO][4090] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:13:55.270859 containerd[1638]: time="2026-03-07T01:13:55.270609048Z" level=info msg="TearDown network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\" successfully" Mar 7 01:13:55.270859 containerd[1638]: time="2026-03-07T01:13:55.270647488Z" level=info msg="StopPodSandbox for \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\" returns successfully" Mar 7 01:13:55.273518 systemd[1]: run-netns-cni\x2de8440e97\x2dc3a8\x2d3938\x2dbb78\x2d090bce3360c5.mount: Deactivated successfully. Mar 7 01:13:55.343177 kubelet[2747]: I0307 01:13:55.341859 2747 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/70b07735-6f2b-40bc-8301-190b364be2aa-nginx-config\") pod \"70b07735-6f2b-40bc-8301-190b364be2aa\" (UID: \"70b07735-6f2b-40bc-8301-190b364be2aa\") " Mar 7 01:13:55.343177 kubelet[2747]: I0307 01:13:55.341936 2747 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngk9n\" (UniqueName: \"kubernetes.io/projected/70b07735-6f2b-40bc-8301-190b364be2aa-kube-api-access-ngk9n\") pod \"70b07735-6f2b-40bc-8301-190b364be2aa\" (UID: \"70b07735-6f2b-40bc-8301-190b364be2aa\") " Mar 7 01:13:55.343177 kubelet[2747]: I0307 01:13:55.341988 2747 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/70b07735-6f2b-40bc-8301-190b364be2aa-whisker-backend-key-pair\") pod \"70b07735-6f2b-40bc-8301-190b364be2aa\" (UID: \"70b07735-6f2b-40bc-8301-190b364be2aa\") " Mar 7 01:13:55.343177 kubelet[2747]: I0307 01:13:55.342015 2747 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70b07735-6f2b-40bc-8301-190b364be2aa-whisker-ca-bundle\") pod \"70b07735-6f2b-40bc-8301-190b364be2aa\" (UID: \"70b07735-6f2b-40bc-8301-190b364be2aa\") " Mar 7 01:13:55.343177 kubelet[2747]: I0307 01:13:55.342658 2747 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70b07735-6f2b-40bc-8301-190b364be2aa-nginx-config" (OuterVolumeSpecName: "nginx-config") pod "70b07735-6f2b-40bc-8301-190b364be2aa" (UID: "70b07735-6f2b-40bc-8301-190b364be2aa"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 7 01:13:55.343496 kubelet[2747]: I0307 01:13:55.342765 2747 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70b07735-6f2b-40bc-8301-190b364be2aa-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "70b07735-6f2b-40bc-8301-190b364be2aa" (UID: "70b07735-6f2b-40bc-8301-190b364be2aa"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 7 01:13:55.350471 kubelet[2747]: I0307 01:13:55.350432 2747 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70b07735-6f2b-40bc-8301-190b364be2aa-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "70b07735-6f2b-40bc-8301-190b364be2aa" (UID: "70b07735-6f2b-40bc-8301-190b364be2aa"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 7 01:13:55.353243 kubelet[2747]: I0307 01:13:55.352679 2747 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70b07735-6f2b-40bc-8301-190b364be2aa-kube-api-access-ngk9n" (OuterVolumeSpecName: "kube-api-access-ngk9n") pod "70b07735-6f2b-40bc-8301-190b364be2aa" (UID: "70b07735-6f2b-40bc-8301-190b364be2aa"). InnerVolumeSpecName "kube-api-access-ngk9n". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 7 01:13:55.354853 systemd[1]: var-lib-kubelet-pods-70b07735\x2d6f2b\x2d40bc\x2d8301\x2d190b364be2aa-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Mar 7 01:13:55.361691 systemd[1]: var-lib-kubelet-pods-70b07735\x2d6f2b\x2d40bc\x2d8301\x2d190b364be2aa-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dngk9n.mount: Deactivated successfully. Mar 7 01:13:55.442520 kubelet[2747]: I0307 01:13:55.442340 2747 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ngk9n\" (UniqueName: \"kubernetes.io/projected/70b07735-6f2b-40bc-8301-190b364be2aa-kube-api-access-ngk9n\") on node \"ci-4081-3-6-n-0c8881e772\" DevicePath \"\"" Mar 7 01:13:55.442520 kubelet[2747]: I0307 01:13:55.442397 2747 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/70b07735-6f2b-40bc-8301-190b364be2aa-whisker-backend-key-pair\") on node \"ci-4081-3-6-n-0c8881e772\" DevicePath \"\"" Mar 7 01:13:55.442520 kubelet[2747]: I0307 01:13:55.442417 2747 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70b07735-6f2b-40bc-8301-190b364be2aa-whisker-ca-bundle\") on node \"ci-4081-3-6-n-0c8881e772\" DevicePath \"\"" Mar 7 01:13:55.442520 kubelet[2747]: I0307 01:13:55.442436 2747 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/70b07735-6f2b-40bc-8301-190b364be2aa-nginx-config\") on node \"ci-4081-3-6-n-0c8881e772\" DevicePath \"\"" Mar 7 01:13:56.186810 kubelet[2747]: I0307 01:13:56.184911 2747 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:13:56.349532 kubelet[2747]: I0307 01:13:56.349487 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/ad5e95f1-b067-4f8a-8bcf-983b76f39373-nginx-config\") pod \"whisker-7fc66dbdd5-dghwg\" (UID: \"ad5e95f1-b067-4f8a-8bcf-983b76f39373\") " pod="calico-system/whisker-7fc66dbdd5-dghwg" Mar 7 01:13:56.349532 kubelet[2747]: I0307 01:13:56.349530 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad5e95f1-b067-4f8a-8bcf-983b76f39373-whisker-ca-bundle\") pod \"whisker-7fc66dbdd5-dghwg\" (UID: \"ad5e95f1-b067-4f8a-8bcf-983b76f39373\") " pod="calico-system/whisker-7fc66dbdd5-dghwg" Mar 7 01:13:56.349984 kubelet[2747]: I0307 01:13:56.349568 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ad5e95f1-b067-4f8a-8bcf-983b76f39373-whisker-backend-key-pair\") pod \"whisker-7fc66dbdd5-dghwg\" (UID: \"ad5e95f1-b067-4f8a-8bcf-983b76f39373\") " pod="calico-system/whisker-7fc66dbdd5-dghwg" Mar 7 01:13:56.349984 kubelet[2747]: I0307 01:13:56.349580 2747 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsf4h\" (UniqueName: \"kubernetes.io/projected/ad5e95f1-b067-4f8a-8bcf-983b76f39373-kube-api-access-tsf4h\") pod \"whisker-7fc66dbdd5-dghwg\" (UID: \"ad5e95f1-b067-4f8a-8bcf-983b76f39373\") " pod="calico-system/whisker-7fc66dbdd5-dghwg" Mar 7 01:13:56.560755 containerd[1638]: time="2026-03-07T01:13:56.560667750Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7fc66dbdd5-dghwg,Uid:ad5e95f1-b067-4f8a-8bcf-983b76f39373,Namespace:calico-system,Attempt:0,}" Mar 7 01:13:56.660235 systemd-networkd[1251]: cali5b3dc730d0d: Link UP Mar 7 01:13:56.661185 systemd-networkd[1251]: cali5b3dc730d0d: Gained carrier Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.592 [ERROR][4208] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.600 [INFO][4208] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0 whisker-7fc66dbdd5- calico-system ad5e95f1-b067-4f8a-8bcf-983b76f39373 917 0 2026-03-07 01:13:56 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:7fc66dbdd5 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4081-3-6-n-0c8881e772 whisker-7fc66dbdd5-dghwg eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali5b3dc730d0d [] [] }} ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Namespace="calico-system" Pod="whisker-7fc66dbdd5-dghwg" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.600 [INFO][4208] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Namespace="calico-system" Pod="whisker-7fc66dbdd5-dghwg" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.622 [INFO][4220] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" HandleID="k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.626 [INFO][4220] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" HandleID="k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002fda50), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-0c8881e772", "pod":"whisker-7fc66dbdd5-dghwg", "timestamp":"2026-03-07 01:13:56.622141587 +0000 UTC"}, Hostname:"ci-4081-3-6-n-0c8881e772", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000548f20)} Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.626 [INFO][4220] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.626 [INFO][4220] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.626 [INFO][4220] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-0c8881e772' Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.628 [INFO][4220] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.632 [INFO][4220] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.635 [INFO][4220] ipam/ipam.go 526: Trying affinity for 192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.637 [INFO][4220] ipam/ipam.go 160: Attempting to load block cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.638 [INFO][4220] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.638 [INFO][4220] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.113.0/26 handle="k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.640 [INFO][4220] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.644 [INFO][4220] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.113.0/26 handle="k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.648 [INFO][4220] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.113.1/26] block=192.168.113.0/26 handle="k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.649 [INFO][4220] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.113.1/26] handle="k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.649 [INFO][4220] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:13:56.673188 containerd[1638]: 2026-03-07 01:13:56.649 [INFO][4220] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.113.1/26] IPv6=[] ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" HandleID="k8s-pod-network.329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" Mar 7 01:13:56.674083 containerd[1638]: 2026-03-07 01:13:56.651 [INFO][4208] cni-plugin/k8s.go 418: Populated endpoint ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Namespace="calico-system" Pod="whisker-7fc66dbdd5-dghwg" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0", GenerateName:"whisker-7fc66dbdd5-", Namespace:"calico-system", SelfLink:"", UID:"ad5e95f1-b067-4f8a-8bcf-983b76f39373", ResourceVersion:"917", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7fc66dbdd5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"", Pod:"whisker-7fc66dbdd5-dghwg", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.113.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali5b3dc730d0d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:13:56.674083 containerd[1638]: 2026-03-07 01:13:56.652 [INFO][4208] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.113.1/32] ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Namespace="calico-system" Pod="whisker-7fc66dbdd5-dghwg" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" Mar 7 01:13:56.674083 containerd[1638]: 2026-03-07 01:13:56.652 [INFO][4208] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5b3dc730d0d ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Namespace="calico-system" Pod="whisker-7fc66dbdd5-dghwg" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" Mar 7 01:13:56.674083 containerd[1638]: 2026-03-07 01:13:56.659 [INFO][4208] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Namespace="calico-system" Pod="whisker-7fc66dbdd5-dghwg" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" Mar 7 01:13:56.674083 containerd[1638]: 2026-03-07 01:13:56.660 [INFO][4208] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Namespace="calico-system" Pod="whisker-7fc66dbdd5-dghwg" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0", GenerateName:"whisker-7fc66dbdd5-", Namespace:"calico-system", SelfLink:"", UID:"ad5e95f1-b067-4f8a-8bcf-983b76f39373", ResourceVersion:"917", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7fc66dbdd5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa", Pod:"whisker-7fc66dbdd5-dghwg", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.113.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali5b3dc730d0d", MAC:"ca:c9:e2:92:ba:e0", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:13:56.674083 containerd[1638]: 2026-03-07 01:13:56.669 [INFO][4208] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa" Namespace="calico-system" Pod="whisker-7fc66dbdd5-dghwg" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--7fc66dbdd5--dghwg-eth0" Mar 7 01:13:56.689795 containerd[1638]: time="2026-03-07T01:13:56.689723303Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:13:56.689795 containerd[1638]: time="2026-03-07T01:13:56.689762503Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:13:56.689795 containerd[1638]: time="2026-03-07T01:13:56.689772413Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:56.690686 containerd[1638]: time="2026-03-07T01:13:56.689839183Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:13:56.774299 containerd[1638]: time="2026-03-07T01:13:56.774257316Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7fc66dbdd5-dghwg,Uid:ad5e95f1-b067-4f8a-8bcf-983b76f39373,Namespace:calico-system,Attempt:0,} returns sandbox id \"329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa\"" Mar 7 01:13:56.776737 containerd[1638]: time="2026-03-07T01:13:56.776711765Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\"" Mar 7 01:13:57.019411 kubelet[2747]: I0307 01:13:57.019366 2747 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70b07735-6f2b-40bc-8301-190b364be2aa" path="/var/lib/kubelet/pods/70b07735-6f2b-40bc-8301-190b364be2aa/volumes" Mar 7 01:13:57.821141 kubelet[2747]: I0307 01:13:57.819650 2747 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:13:57.932361 systemd[1]: run-containerd-runc-k8s.io-4d6f2b013a064a1d00774969e13f724b0eda1d044831eed38b7c44ff377ebbc2-runc.hiAPCu.mount: Deactivated successfully. Mar 7 01:13:58.146000 systemd-networkd[1251]: cali5b3dc730d0d: Gained IPv6LL Mar 7 01:13:58.400144 containerd[1638]: time="2026-03-07T01:13:58.399883698Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:58.400749 containerd[1638]: time="2026-03-07T01:13:58.400577988Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.4: active requests=0, bytes read=6039889" Mar 7 01:13:58.401576 containerd[1638]: time="2026-03-07T01:13:58.401465228Z" level=info msg="ImageCreate event name:\"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:58.403264 containerd[1638]: time="2026-03-07T01:13:58.403232518Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:13:58.403929 containerd[1638]: time="2026-03-07T01:13:58.403757997Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.4\" with image id \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\", size \"7595926\" in 1.627019211s" Mar 7 01:13:58.403929 containerd[1638]: time="2026-03-07T01:13:58.403790257Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\" returns image reference \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\"" Mar 7 01:13:58.407344 containerd[1638]: time="2026-03-07T01:13:58.407314667Z" level=info msg="CreateContainer within sandbox \"329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Mar 7 01:13:58.427758 containerd[1638]: time="2026-03-07T01:13:58.427695454Z" level=info msg="CreateContainer within sandbox \"329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"6cd0b83fb1a231b658f744e6ec1601891a1b3d2b73dbd85df2eecbbbf3e79541\"" Mar 7 01:13:58.428297 containerd[1638]: time="2026-03-07T01:13:58.428271384Z" level=info msg="StartContainer for \"6cd0b83fb1a231b658f744e6ec1601891a1b3d2b73dbd85df2eecbbbf3e79541\"" Mar 7 01:13:58.484869 containerd[1638]: time="2026-03-07T01:13:58.484777207Z" level=info msg="StartContainer for \"6cd0b83fb1a231b658f744e6ec1601891a1b3d2b73dbd85df2eecbbbf3e79541\" returns successfully" Mar 7 01:13:58.487259 containerd[1638]: time="2026-03-07T01:13:58.487030837Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\"" Mar 7 01:13:59.254884 kubelet[2747]: I0307 01:13:59.254461 2747 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:14:00.029605 kernel: calico-node[4423]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Mar 7 01:14:00.457413 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3194058778.mount: Deactivated successfully. Mar 7 01:14:00.472154 containerd[1638]: time="2026-03-07T01:14:00.471588891Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:00.473038 containerd[1638]: time="2026-03-07T01:14:00.473011772Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.4: active requests=0, bytes read=17609475" Mar 7 01:14:00.474232 containerd[1638]: time="2026-03-07T01:14:00.474216551Z" level=info msg="ImageCreate event name:\"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:00.476139 containerd[1638]: time="2026-03-07T01:14:00.476123661Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:00.477207 containerd[1638]: time="2026-03-07T01:14:00.477121502Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" with image id \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\", size \"17609305\" in 1.990072035s" Mar 7 01:14:00.477207 containerd[1638]: time="2026-03-07T01:14:00.477142662Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" returns image reference \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\"" Mar 7 01:14:00.482558 containerd[1638]: time="2026-03-07T01:14:00.482423491Z" level=info msg="CreateContainer within sandbox \"329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Mar 7 01:14:00.494728 containerd[1638]: time="2026-03-07T01:14:00.494708030Z" level=info msg="CreateContainer within sandbox \"329e79497d3fe1909bb3444aba734976acc583f2ed27ddd666aa91593c47d2fa\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"4b1aaa8482249dcd33a753d40d148051893ea22868c1c10a54bdef28fcf3fa58\"" Mar 7 01:14:00.495855 containerd[1638]: time="2026-03-07T01:14:00.495135470Z" level=info msg="StartContainer for \"4b1aaa8482249dcd33a753d40d148051893ea22868c1c10a54bdef28fcf3fa58\"" Mar 7 01:14:00.570210 containerd[1638]: time="2026-03-07T01:14:00.570176295Z" level=info msg="StartContainer for \"4b1aaa8482249dcd33a753d40d148051893ea22868c1c10a54bdef28fcf3fa58\" returns successfully" Mar 7 01:14:00.633829 systemd-networkd[1251]: vxlan.calico: Link UP Mar 7 01:14:00.633849 systemd-networkd[1251]: vxlan.calico: Gained carrier Mar 7 01:14:02.180665 systemd-networkd[1251]: vxlan.calico: Gained IPv6LL Mar 7 01:14:05.021647 containerd[1638]: time="2026-03-07T01:14:05.020665881Z" level=info msg="StopPodSandbox for \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\"" Mar 7 01:14:05.022536 containerd[1638]: time="2026-03-07T01:14:05.020831231Z" level=info msg="StopPodSandbox for \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\"" Mar 7 01:14:05.099601 kubelet[2747]: I0307 01:14:05.099322 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-7fc66dbdd5-dghwg" podStartSLOduration=5.397485642 podStartE2EDuration="9.099302507s" podCreationTimestamp="2026-03-07 01:13:56 +0000 UTC" firstStartedPulling="2026-03-07 01:13:56.775877256 +0000 UTC m=+31.842788983" lastFinishedPulling="2026-03-07 01:14:00.477694121 +0000 UTC m=+35.544605848" observedRunningTime="2026-03-07 01:14:01.218681714 +0000 UTC m=+36.285593461" watchObservedRunningTime="2026-03-07 01:14:05.099302507 +0000 UTC m=+40.166214234" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.100 [INFO][4634] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.100 [INFO][4634] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" iface="eth0" netns="/var/run/netns/cni-3a1d765b-fe78-bb9e-1381-bb07d237b676" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.100 [INFO][4634] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" iface="eth0" netns="/var/run/netns/cni-3a1d765b-fe78-bb9e-1381-bb07d237b676" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.101 [INFO][4634] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" iface="eth0" netns="/var/run/netns/cni-3a1d765b-fe78-bb9e-1381-bb07d237b676" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.101 [INFO][4634] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.101 [INFO][4634] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.124 [INFO][4648] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.124 [INFO][4648] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.124 [INFO][4648] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.129 [WARNING][4648] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.129 [INFO][4648] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.130 [INFO][4648] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:05.133861 containerd[1638]: 2026-03-07 01:14:05.132 [INFO][4634] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:05.135690 containerd[1638]: time="2026-03-07T01:14:05.135589830Z" level=info msg="TearDown network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\" successfully" Mar 7 01:14:05.135690 containerd[1638]: time="2026-03-07T01:14:05.135614030Z" level=info msg="StopPodSandbox for \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\" returns successfully" Mar 7 01:14:05.137557 systemd[1]: run-netns-cni\x2d3a1d765b\x2dfe78\x2dbb9e\x2d1381\x2dbb07d237b676.mount: Deactivated successfully. Mar 7 01:14:05.138597 containerd[1638]: time="2026-03-07T01:14:05.137862320Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-689889c887-ntpwf,Uid:a9f53e47-46dd-420a-9df1-28f9cae95a81,Namespace:calico-system,Attempt:1,}" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.109 [INFO][4635] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.109 [INFO][4635] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" iface="eth0" netns="/var/run/netns/cni-aecfd41d-0cb4-c2e8-e96d-acdebc641a7b" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.109 [INFO][4635] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" iface="eth0" netns="/var/run/netns/cni-aecfd41d-0cb4-c2e8-e96d-acdebc641a7b" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.109 [INFO][4635] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" iface="eth0" netns="/var/run/netns/cni-aecfd41d-0cb4-c2e8-e96d-acdebc641a7b" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.109 [INFO][4635] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.109 [INFO][4635] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.126 [INFO][4654] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.126 [INFO][4654] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.130 [INFO][4654] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.139 [WARNING][4654] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.139 [INFO][4654] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.140 [INFO][4654] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:05.145056 containerd[1638]: 2026-03-07 01:14:05.142 [INFO][4635] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:05.145341 containerd[1638]: time="2026-03-07T01:14:05.145252050Z" level=info msg="TearDown network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\" successfully" Mar 7 01:14:05.145341 containerd[1638]: time="2026-03-07T01:14:05.145272860Z" level=info msg="StopPodSandbox for \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\" returns successfully" Mar 7 01:14:05.148072 containerd[1638]: time="2026-03-07T01:14:05.147819480Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-9qlr7,Uid:ca853831-8f8c-4ad5-a296-b442696cf9d1,Namespace:calico-system,Attempt:1,}" Mar 7 01:14:05.148135 systemd[1]: run-netns-cni\x2daecfd41d\x2d0cb4\x2dc2e8\x2de96d\x2dacdebc641a7b.mount: Deactivated successfully. Mar 7 01:14:05.265451 systemd-networkd[1251]: cali42e2d916126: Link UP Mar 7 01:14:05.266184 systemd-networkd[1251]: cali42e2d916126: Gained carrier Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.191 [INFO][4663] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0 calico-apiserver-689889c887- calico-system a9f53e47-46dd-420a-9df1-28f9cae95a81 963 0 2026-03-07 01:13:39 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:689889c887 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081-3-6-n-0c8881e772 calico-apiserver-689889c887-ntpwf eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali42e2d916126 [] [] }} ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Namespace="calico-system" Pod="calico-apiserver-689889c887-ntpwf" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.191 [INFO][4663] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Namespace="calico-system" Pod="calico-apiserver-689889c887-ntpwf" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.217 [INFO][4686] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" HandleID="k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.223 [INFO][4686] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" HandleID="k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002fd7a0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-0c8881e772", "pod":"calico-apiserver-689889c887-ntpwf", "timestamp":"2026-03-07 01:14:05.217908225 +0000 UTC"}, Hostname:"ci-4081-3-6-n-0c8881e772", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0003a3760)} Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.223 [INFO][4686] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.223 [INFO][4686] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.223 [INFO][4686] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-0c8881e772' Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.225 [INFO][4686] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.230 [INFO][4686] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.239 [INFO][4686] ipam/ipam.go 526: Trying affinity for 192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.241 [INFO][4686] ipam/ipam.go 160: Attempting to load block cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.243 [INFO][4686] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.243 [INFO][4686] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.113.0/26 handle="k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.245 [INFO][4686] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587 Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.251 [INFO][4686] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.113.0/26 handle="k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.256 [INFO][4686] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.113.2/26] block=192.168.113.0/26 handle="k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.257 [INFO][4686] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.113.2/26] handle="k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.257 [INFO][4686] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:05.279994 containerd[1638]: 2026-03-07 01:14:05.257 [INFO][4686] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.113.2/26] IPv6=[] ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" HandleID="k8s-pod-network.4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.281419 containerd[1638]: 2026-03-07 01:14:05.262 [INFO][4663] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Namespace="calico-system" Pod="calico-apiserver-689889c887-ntpwf" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0", GenerateName:"calico-apiserver-689889c887-", Namespace:"calico-system", SelfLink:"", UID:"a9f53e47-46dd-420a-9df1-28f9cae95a81", ResourceVersion:"963", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"689889c887", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"", Pod:"calico-apiserver-689889c887-ntpwf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.113.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali42e2d916126", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:05.281419 containerd[1638]: 2026-03-07 01:14:05.263 [INFO][4663] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.113.2/32] ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Namespace="calico-system" Pod="calico-apiserver-689889c887-ntpwf" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.281419 containerd[1638]: 2026-03-07 01:14:05.263 [INFO][4663] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali42e2d916126 ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Namespace="calico-system" Pod="calico-apiserver-689889c887-ntpwf" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.281419 containerd[1638]: 2026-03-07 01:14:05.266 [INFO][4663] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Namespace="calico-system" Pod="calico-apiserver-689889c887-ntpwf" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.281419 containerd[1638]: 2026-03-07 01:14:05.266 [INFO][4663] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Namespace="calico-system" Pod="calico-apiserver-689889c887-ntpwf" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0", GenerateName:"calico-apiserver-689889c887-", Namespace:"calico-system", SelfLink:"", UID:"a9f53e47-46dd-420a-9df1-28f9cae95a81", ResourceVersion:"963", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"689889c887", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587", Pod:"calico-apiserver-689889c887-ntpwf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.113.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali42e2d916126", MAC:"f6:93:1b:1d:f9:98", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:05.281419 containerd[1638]: 2026-03-07 01:14:05.274 [INFO][4663] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587" Namespace="calico-system" Pod="calico-apiserver-689889c887-ntpwf" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:05.307580 containerd[1638]: time="2026-03-07T01:14:05.307497841Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:14:05.308140 containerd[1638]: time="2026-03-07T01:14:05.308103292Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:14:05.308574 containerd[1638]: time="2026-03-07T01:14:05.308213602Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:05.308574 containerd[1638]: time="2026-03-07T01:14:05.308293092Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:05.361240 systemd-networkd[1251]: calidb8eca2181a: Link UP Mar 7 01:14:05.361387 systemd-networkd[1251]: calidb8eca2181a: Gained carrier Mar 7 01:14:05.371910 containerd[1638]: time="2026-03-07T01:14:05.371776906Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-689889c887-ntpwf,Uid:a9f53e47-46dd-420a-9df1-28f9cae95a81,Namespace:calico-system,Attempt:1,} returns sandbox id \"4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587\"" Mar 7 01:14:05.375588 containerd[1638]: time="2026-03-07T01:14:05.375472077Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.204 [INFO][4673] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0 goldmane-5b85766d88- calico-system ca853831-8f8c-4ad5-a296-b442696cf9d1 964 0 2026-03-07 01:13:39 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:5b85766d88 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4081-3-6-n-0c8881e772 goldmane-5b85766d88-9qlr7 eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calidb8eca2181a [] [] }} ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Namespace="calico-system" Pod="goldmane-5b85766d88-9qlr7" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.204 [INFO][4673] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Namespace="calico-system" Pod="goldmane-5b85766d88-9qlr7" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.240 [INFO][4693] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" HandleID="k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.250 [INFO][4693] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" HandleID="k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000277940), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-0c8881e772", "pod":"goldmane-5b85766d88-9qlr7", "timestamp":"2026-03-07 01:14:05.240563117 +0000 UTC"}, Hostname:"ci-4081-3-6-n-0c8881e772", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0003b4f20)} Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.251 [INFO][4693] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.257 [INFO][4693] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.257 [INFO][4693] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-0c8881e772' Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.327 [INFO][4693] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.331 [INFO][4693] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.336 [INFO][4693] ipam/ipam.go 526: Trying affinity for 192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.338 [INFO][4693] ipam/ipam.go 160: Attempting to load block cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.339 [INFO][4693] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.339 [INFO][4693] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.113.0/26 handle="k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.342 [INFO][4693] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042 Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.345 [INFO][4693] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.113.0/26 handle="k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.350 [INFO][4693] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.113.3/26] block=192.168.113.0/26 handle="k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.350 [INFO][4693] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.113.3/26] handle="k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.350 [INFO][4693] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:05.382527 containerd[1638]: 2026-03-07 01:14:05.350 [INFO][4693] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.113.3/26] IPv6=[] ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" HandleID="k8s-pod-network.1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.383663 containerd[1638]: 2026-03-07 01:14:05.356 [INFO][4673] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Namespace="calico-system" Pod="goldmane-5b85766d88-9qlr7" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"ca853831-8f8c-4ad5-a296-b442696cf9d1", ResourceVersion:"964", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"", Pod:"goldmane-5b85766d88-9qlr7", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.113.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calidb8eca2181a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:05.383663 containerd[1638]: 2026-03-07 01:14:05.357 [INFO][4673] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.113.3/32] ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Namespace="calico-system" Pod="goldmane-5b85766d88-9qlr7" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.383663 containerd[1638]: 2026-03-07 01:14:05.358 [INFO][4673] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidb8eca2181a ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Namespace="calico-system" Pod="goldmane-5b85766d88-9qlr7" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.383663 containerd[1638]: 2026-03-07 01:14:05.360 [INFO][4673] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Namespace="calico-system" Pod="goldmane-5b85766d88-9qlr7" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.383663 containerd[1638]: 2026-03-07 01:14:05.361 [INFO][4673] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Namespace="calico-system" Pod="goldmane-5b85766d88-9qlr7" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"ca853831-8f8c-4ad5-a296-b442696cf9d1", ResourceVersion:"964", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042", Pod:"goldmane-5b85766d88-9qlr7", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.113.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calidb8eca2181a", MAC:"12:2f:8f:d0:88:c1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:05.383663 containerd[1638]: 2026-03-07 01:14:05.372 [INFO][4673] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042" Namespace="calico-system" Pod="goldmane-5b85766d88-9qlr7" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:05.401029 containerd[1638]: time="2026-03-07T01:14:05.400960228Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:14:05.401350 containerd[1638]: time="2026-03-07T01:14:05.401233039Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:14:05.401350 containerd[1638]: time="2026-03-07T01:14:05.401261729Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:05.401411 containerd[1638]: time="2026-03-07T01:14:05.401341939Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:05.457256 containerd[1638]: time="2026-03-07T01:14:05.457217973Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-9qlr7,Uid:ca853831-8f8c-4ad5-a296-b442696cf9d1,Namespace:calico-system,Attempt:1,} returns sandbox id \"1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042\"" Mar 7 01:14:06.787881 systemd-networkd[1251]: calidb8eca2181a: Gained IPv6LL Mar 7 01:14:06.914126 systemd-networkd[1251]: cali42e2d916126: Gained IPv6LL Mar 7 01:14:07.017629 containerd[1638]: time="2026-03-07T01:14:07.017569465Z" level=info msg="StopPodSandbox for \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\"" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.064 [INFO][4859] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.064 [INFO][4859] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" iface="eth0" netns="/var/run/netns/cni-e587fe0d-8d0e-d4b0-f31e-7059c3402257" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.064 [INFO][4859] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" iface="eth0" netns="/var/run/netns/cni-e587fe0d-8d0e-d4b0-f31e-7059c3402257" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.064 [INFO][4859] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" iface="eth0" netns="/var/run/netns/cni-e587fe0d-8d0e-d4b0-f31e-7059c3402257" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.064 [INFO][4859] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.064 [INFO][4859] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.109 [INFO][4866] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.109 [INFO][4866] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.109 [INFO][4866] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.117 [WARNING][4866] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.117 [INFO][4866] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.118 [INFO][4866] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:07.129001 containerd[1638]: 2026-03-07 01:14:07.122 [INFO][4859] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:07.129001 containerd[1638]: time="2026-03-07T01:14:07.127038097Z" level=info msg="TearDown network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\" successfully" Mar 7 01:14:07.129001 containerd[1638]: time="2026-03-07T01:14:07.127066107Z" level=info msg="StopPodSandbox for \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\" returns successfully" Mar 7 01:14:07.129498 containerd[1638]: time="2026-03-07T01:14:07.129104187Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5dc64846bc-k762g,Uid:7f399497-ed30-4f43-b4f2-62919a08223a,Namespace:calico-system,Attempt:1,}" Mar 7 01:14:07.132294 systemd[1]: run-netns-cni\x2de587fe0d\x2d8d0e\x2dd4b0\x2df31e\x2d7059c3402257.mount: Deactivated successfully. Mar 7 01:14:07.277456 systemd-networkd[1251]: cali295ecf6c436: Link UP Mar 7 01:14:07.278836 systemd-networkd[1251]: cali295ecf6c436: Gained carrier Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.201 [INFO][4872] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0 calico-kube-controllers-5dc64846bc- calico-system 7f399497-ed30-4f43-b4f2-62919a08223a 978 0 2026-03-07 01:13:40 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5dc64846bc projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081-3-6-n-0c8881e772 calico-kube-controllers-5dc64846bc-k762g eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali295ecf6c436 [] [] }} ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Namespace="calico-system" Pod="calico-kube-controllers-5dc64846bc-k762g" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.201 [INFO][4872] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Namespace="calico-system" Pod="calico-kube-controllers-5dc64846bc-k762g" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.237 [INFO][4884] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" HandleID="k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.246 [INFO][4884] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" HandleID="k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002fde80), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-0c8881e772", "pod":"calico-kube-controllers-5dc64846bc-k762g", "timestamp":"2026-03-07 01:14:07.23728645 +0000 UTC"}, Hostname:"ci-4081-3-6-n-0c8881e772", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0001862c0)} Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.246 [INFO][4884] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.246 [INFO][4884] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.246 [INFO][4884] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-0c8881e772' Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.248 [INFO][4884] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.252 [INFO][4884] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.256 [INFO][4884] ipam/ipam.go 526: Trying affinity for 192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.257 [INFO][4884] ipam/ipam.go 160: Attempting to load block cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.259 [INFO][4884] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.259 [INFO][4884] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.113.0/26 handle="k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.261 [INFO][4884] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.264 [INFO][4884] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.113.0/26 handle="k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.269 [INFO][4884] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.113.4/26] block=192.168.113.0/26 handle="k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.269 [INFO][4884] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.113.4/26] handle="k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.269 [INFO][4884] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:07.295724 containerd[1638]: 2026-03-07 01:14:07.269 [INFO][4884] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.113.4/26] IPv6=[] ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" HandleID="k8s-pod-network.40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.297146 containerd[1638]: 2026-03-07 01:14:07.274 [INFO][4872] cni-plugin/k8s.go 418: Populated endpoint ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Namespace="calico-system" Pod="calico-kube-controllers-5dc64846bc-k762g" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0", GenerateName:"calico-kube-controllers-5dc64846bc-", Namespace:"calico-system", SelfLink:"", UID:"7f399497-ed30-4f43-b4f2-62919a08223a", ResourceVersion:"978", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5dc64846bc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"", Pod:"calico-kube-controllers-5dc64846bc-k762g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.113.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali295ecf6c436", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:07.297146 containerd[1638]: 2026-03-07 01:14:07.274 [INFO][4872] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.113.4/32] ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Namespace="calico-system" Pod="calico-kube-controllers-5dc64846bc-k762g" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.297146 containerd[1638]: 2026-03-07 01:14:07.274 [INFO][4872] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali295ecf6c436 ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Namespace="calico-system" Pod="calico-kube-controllers-5dc64846bc-k762g" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.297146 containerd[1638]: 2026-03-07 01:14:07.280 [INFO][4872] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Namespace="calico-system" Pod="calico-kube-controllers-5dc64846bc-k762g" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.297146 containerd[1638]: 2026-03-07 01:14:07.280 [INFO][4872] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Namespace="calico-system" Pod="calico-kube-controllers-5dc64846bc-k762g" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0", GenerateName:"calico-kube-controllers-5dc64846bc-", Namespace:"calico-system", SelfLink:"", UID:"7f399497-ed30-4f43-b4f2-62919a08223a", ResourceVersion:"978", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5dc64846bc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa", Pod:"calico-kube-controllers-5dc64846bc-k762g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.113.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali295ecf6c436", MAC:"6e:fd:90:0b:cb:aa", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:07.297146 containerd[1638]: 2026-03-07 01:14:07.292 [INFO][4872] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa" Namespace="calico-system" Pod="calico-kube-controllers-5dc64846bc-k762g" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:07.335861 containerd[1638]: time="2026-03-07T01:14:07.335338861Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:14:07.337458 containerd[1638]: time="2026-03-07T01:14:07.336089781Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:14:07.337458 containerd[1638]: time="2026-03-07T01:14:07.336103721Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:07.337458 containerd[1638]: time="2026-03-07T01:14:07.336186231Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:07.447453 containerd[1638]: time="2026-03-07T01:14:07.447413174Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5dc64846bc-k762g,Uid:7f399497-ed30-4f43-b4f2-62919a08223a,Namespace:calico-system,Attempt:1,} returns sandbox id \"40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa\"" Mar 7 01:14:07.740015 containerd[1638]: time="2026-03-07T01:14:07.739713966Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:07.740755 containerd[1638]: time="2026-03-07T01:14:07.740565107Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=48415780" Mar 7 01:14:07.741741 containerd[1638]: time="2026-03-07T01:14:07.741694027Z" level=info msg="ImageCreate event name:\"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:07.744572 containerd[1638]: time="2026-03-07T01:14:07.743684067Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:07.744572 containerd[1638]: time="2026-03-07T01:14:07.744411677Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"49971841\" in 2.36891678s" Mar 7 01:14:07.744572 containerd[1638]: time="2026-03-07T01:14:07.744434077Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\"" Mar 7 01:14:07.746662 containerd[1638]: time="2026-03-07T01:14:07.746640297Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\"" Mar 7 01:14:07.749313 containerd[1638]: time="2026-03-07T01:14:07.749286317Z" level=info msg="CreateContainer within sandbox \"4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 7 01:14:07.759456 containerd[1638]: time="2026-03-07T01:14:07.759422779Z" level=info msg="CreateContainer within sandbox \"4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"618ee0cf4dc4be12668d70768d3e22c6af689abaffd0bcf40e6873a071419adc\"" Mar 7 01:14:07.759898 containerd[1638]: time="2026-03-07T01:14:07.759868139Z" level=info msg="StartContainer for \"618ee0cf4dc4be12668d70768d3e22c6af689abaffd0bcf40e6873a071419adc\"" Mar 7 01:14:07.820492 containerd[1638]: time="2026-03-07T01:14:07.820413786Z" level=info msg="StartContainer for \"618ee0cf4dc4be12668d70768d3e22c6af689abaffd0bcf40e6873a071419adc\" returns successfully" Mar 7 01:14:08.018510 containerd[1638]: time="2026-03-07T01:14:08.018376459Z" level=info msg="StopPodSandbox for \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\"" Mar 7 01:14:08.020558 containerd[1638]: time="2026-03-07T01:14:08.020014068Z" level=info msg="StopPodSandbox for \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\"" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.073 [INFO][5012] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.074 [INFO][5012] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" iface="eth0" netns="/var/run/netns/cni-394c1624-77d1-ba51-8d35-507b137d8c3c" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.074 [INFO][5012] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" iface="eth0" netns="/var/run/netns/cni-394c1624-77d1-ba51-8d35-507b137d8c3c" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.074 [INFO][5012] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" iface="eth0" netns="/var/run/netns/cni-394c1624-77d1-ba51-8d35-507b137d8c3c" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.074 [INFO][5012] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.074 [INFO][5012] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.093 [INFO][5023] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.093 [INFO][5023] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.093 [INFO][5023] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.097 [WARNING][5023] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.098 [INFO][5023] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.099 [INFO][5023] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:08.104156 containerd[1638]: 2026-03-07 01:14:08.102 [INFO][5012] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:08.104780 containerd[1638]: time="2026-03-07T01:14:08.104727929Z" level=info msg="TearDown network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\" successfully" Mar 7 01:14:08.104780 containerd[1638]: time="2026-03-07T01:14:08.104750759Z" level=info msg="StopPodSandbox for \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\" returns successfully" Mar 7 01:14:08.105602 containerd[1638]: time="2026-03-07T01:14:08.105354220Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-4xvx5,Uid:99f62e43-c05f-4b23-90d0-600585809c83,Namespace:kube-system,Attempt:1,}" Mar 7 01:14:08.150591 systemd[1]: run-netns-cni\x2d394c1624\x2d77d1\x2dba51\x2d8d35\x2d507b137d8c3c.mount: Deactivated successfully. Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.123 [INFO][5008] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.123 [INFO][5008] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" iface="eth0" netns="/var/run/netns/cni-0a3ce49b-b3b9-d33e-eb58-a6f46d8bbdcd" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.124 [INFO][5008] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" iface="eth0" netns="/var/run/netns/cni-0a3ce49b-b3b9-d33e-eb58-a6f46d8bbdcd" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.124 [INFO][5008] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" iface="eth0" netns="/var/run/netns/cni-0a3ce49b-b3b9-d33e-eb58-a6f46d8bbdcd" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.124 [INFO][5008] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.124 [INFO][5008] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.175 [INFO][5034] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.175 [INFO][5034] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.175 [INFO][5034] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.180 [WARNING][5034] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.180 [INFO][5034] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.182 [INFO][5034] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:08.202156 containerd[1638]: 2026-03-07 01:14:08.184 [INFO][5008] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:08.204034 containerd[1638]: time="2026-03-07T01:14:08.203965363Z" level=info msg="TearDown network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\" successfully" Mar 7 01:14:08.204111 containerd[1638]: time="2026-03-07T01:14:08.204093973Z" level=info msg="StopPodSandbox for \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\" returns successfully" Mar 7 01:14:08.206004 systemd[1]: run-netns-cni\x2d0a3ce49b\x2db3b9\x2dd33e\x2deb58\x2da6f46d8bbdcd.mount: Deactivated successfully. Mar 7 01:14:08.208959 containerd[1638]: time="2026-03-07T01:14:08.208911144Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-98kpm,Uid:1d72866e-1f81-4c34-a51f-96e1e7fc603a,Namespace:kube-system,Attempt:1,}" Mar 7 01:14:08.271772 kubelet[2747]: I0307 01:14:08.269480 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-689889c887-ntpwf" podStartSLOduration=26.896908141 podStartE2EDuration="29.269464372s" podCreationTimestamp="2026-03-07 01:13:39 +0000 UTC" firstStartedPulling="2026-03-07 01:14:05.372753607 +0000 UTC m=+40.439665324" lastFinishedPulling="2026-03-07 01:14:07.745309838 +0000 UTC m=+42.812221555" observedRunningTime="2026-03-07 01:14:08.267755321 +0000 UTC m=+43.334667058" watchObservedRunningTime="2026-03-07 01:14:08.269464372 +0000 UTC m=+43.336376099" Mar 7 01:14:08.298791 systemd-networkd[1251]: calib83eb51f121: Link UP Mar 7 01:14:08.300847 systemd-networkd[1251]: calib83eb51f121: Gained carrier Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.165 [INFO][5031] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0 coredns-674b8bbfcf- kube-system 99f62e43-c05f-4b23-90d0-600585809c83 989 0 2026-03-07 01:13:30 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081-3-6-n-0c8881e772 coredns-674b8bbfcf-4xvx5 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calib83eb51f121 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Namespace="kube-system" Pod="coredns-674b8bbfcf-4xvx5" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.165 [INFO][5031] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Namespace="kube-system" Pod="coredns-674b8bbfcf-4xvx5" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.223 [INFO][5048] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" HandleID="k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.245 [INFO][5048] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" HandleID="k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003c67c0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081-3-6-n-0c8881e772", "pod":"coredns-674b8bbfcf-4xvx5", "timestamp":"2026-03-07 01:14:08.223591536 +0000 UTC"}, Hostname:"ci-4081-3-6-n-0c8881e772", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000470c60)} Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.245 [INFO][5048] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.245 [INFO][5048] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.245 [INFO][5048] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-0c8881e772' Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.247 [INFO][5048] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.252 [INFO][5048] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.260 [INFO][5048] ipam/ipam.go 526: Trying affinity for 192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.264 [INFO][5048] ipam/ipam.go 160: Attempting to load block cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.273 [INFO][5048] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.273 [INFO][5048] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.113.0/26 handle="k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.275 [INFO][5048] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.279 [INFO][5048] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.113.0/26 handle="k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.283 [INFO][5048] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.113.5/26] block=192.168.113.0/26 handle="k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.283 [INFO][5048] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.113.5/26] handle="k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.284 [INFO][5048] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:08.325738 containerd[1638]: 2026-03-07 01:14:08.284 [INFO][5048] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.113.5/26] IPv6=[] ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" HandleID="k8s-pod-network.66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.326188 containerd[1638]: 2026-03-07 01:14:08.290 [INFO][5031] cni-plugin/k8s.go 418: Populated endpoint ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Namespace="kube-system" Pod="coredns-674b8bbfcf-4xvx5" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"99f62e43-c05f-4b23-90d0-600585809c83", ResourceVersion:"989", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"", Pod:"coredns-674b8bbfcf-4xvx5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.113.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calib83eb51f121", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:08.326188 containerd[1638]: 2026-03-07 01:14:08.290 [INFO][5031] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.113.5/32] ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Namespace="kube-system" Pod="coredns-674b8bbfcf-4xvx5" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.326188 containerd[1638]: 2026-03-07 01:14:08.290 [INFO][5031] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib83eb51f121 ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Namespace="kube-system" Pod="coredns-674b8bbfcf-4xvx5" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.326188 containerd[1638]: 2026-03-07 01:14:08.302 [INFO][5031] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Namespace="kube-system" Pod="coredns-674b8bbfcf-4xvx5" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.326188 containerd[1638]: 2026-03-07 01:14:08.303 [INFO][5031] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Namespace="kube-system" Pod="coredns-674b8bbfcf-4xvx5" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"99f62e43-c05f-4b23-90d0-600585809c83", ResourceVersion:"989", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b", Pod:"coredns-674b8bbfcf-4xvx5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.113.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calib83eb51f121", MAC:"12:47:6c:1c:df:08", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:08.326188 containerd[1638]: 2026-03-07 01:14:08.319 [INFO][5031] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b" Namespace="kube-system" Pod="coredns-674b8bbfcf-4xvx5" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:08.421068 containerd[1638]: time="2026-03-07T01:14:08.420382422Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:14:08.421068 containerd[1638]: time="2026-03-07T01:14:08.420438262Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:14:08.421068 containerd[1638]: time="2026-03-07T01:14:08.420448342Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:08.421068 containerd[1638]: time="2026-03-07T01:14:08.420528552Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:08.445207 systemd-networkd[1251]: calif37b16151c8: Link UP Mar 7 01:14:08.447183 systemd-networkd[1251]: calif37b16151c8: Gained carrier Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.323 [INFO][5055] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0 coredns-674b8bbfcf- kube-system 1d72866e-1f81-4c34-a51f-96e1e7fc603a 990 0 2026-03-07 01:13:30 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081-3-6-n-0c8881e772 coredns-674b8bbfcf-98kpm eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calif37b16151c8 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Namespace="kube-system" Pod="coredns-674b8bbfcf-98kpm" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.323 [INFO][5055] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Namespace="kube-system" Pod="coredns-674b8bbfcf-98kpm" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.383 [INFO][5077] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" HandleID="k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.393 [INFO][5077] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" HandleID="k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fe90), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081-3-6-n-0c8881e772", "pod":"coredns-674b8bbfcf-98kpm", "timestamp":"2026-03-07 01:14:08.383160816 +0000 UTC"}, Hostname:"ci-4081-3-6-n-0c8881e772", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000186580)} Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.393 [INFO][5077] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.393 [INFO][5077] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.393 [INFO][5077] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-0c8881e772' Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.395 [INFO][5077] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.399 [INFO][5077] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.403 [INFO][5077] ipam/ipam.go 526: Trying affinity for 192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.407 [INFO][5077] ipam/ipam.go 160: Attempting to load block cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.409 [INFO][5077] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.409 [INFO][5077] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.113.0/26 handle="k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.411 [INFO][5077] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7 Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.421 [INFO][5077] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.113.0/26 handle="k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.428 [INFO][5077] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.113.6/26] block=192.168.113.0/26 handle="k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.428 [INFO][5077] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.113.6/26] handle="k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.428 [INFO][5077] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:08.473302 containerd[1638]: 2026-03-07 01:14:08.428 [INFO][5077] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.113.6/26] IPv6=[] ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" HandleID="k8s-pod-network.4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.473751 containerd[1638]: 2026-03-07 01:14:08.436 [INFO][5055] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Namespace="kube-system" Pod="coredns-674b8bbfcf-98kpm" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"1d72866e-1f81-4c34-a51f-96e1e7fc603a", ResourceVersion:"990", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"", Pod:"coredns-674b8bbfcf-98kpm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.113.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif37b16151c8", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:08.473751 containerd[1638]: 2026-03-07 01:14:08.436 [INFO][5055] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.113.6/32] ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Namespace="kube-system" Pod="coredns-674b8bbfcf-98kpm" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.473751 containerd[1638]: 2026-03-07 01:14:08.436 [INFO][5055] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif37b16151c8 ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Namespace="kube-system" Pod="coredns-674b8bbfcf-98kpm" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.473751 containerd[1638]: 2026-03-07 01:14:08.446 [INFO][5055] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Namespace="kube-system" Pod="coredns-674b8bbfcf-98kpm" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.473751 containerd[1638]: 2026-03-07 01:14:08.456 [INFO][5055] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Namespace="kube-system" Pod="coredns-674b8bbfcf-98kpm" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"1d72866e-1f81-4c34-a51f-96e1e7fc603a", ResourceVersion:"990", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7", Pod:"coredns-674b8bbfcf-98kpm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.113.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif37b16151c8", MAC:"06:52:54:35:9c:6a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:08.473751 containerd[1638]: 2026-03-07 01:14:08.467 [INFO][5055] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7" Namespace="kube-system" Pod="coredns-674b8bbfcf-98kpm" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:08.516311 containerd[1638]: time="2026-03-07T01:14:08.515932164Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:14:08.516311 containerd[1638]: time="2026-03-07T01:14:08.516003684Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:14:08.516311 containerd[1638]: time="2026-03-07T01:14:08.516018984Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:08.516311 containerd[1638]: time="2026-03-07T01:14:08.516114314Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:08.577337 containerd[1638]: time="2026-03-07T01:14:08.577311772Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-4xvx5,Uid:99f62e43-c05f-4b23-90d0-600585809c83,Namespace:kube-system,Attempt:1,} returns sandbox id \"66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b\"" Mar 7 01:14:08.609328 containerd[1638]: time="2026-03-07T01:14:08.609250627Z" level=info msg="CreateContainer within sandbox \"66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 7 01:14:08.621458 containerd[1638]: time="2026-03-07T01:14:08.621418558Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-98kpm,Uid:1d72866e-1f81-4c34-a51f-96e1e7fc603a,Namespace:kube-system,Attempt:1,} returns sandbox id \"4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7\"" Mar 7 01:14:08.624815 containerd[1638]: time="2026-03-07T01:14:08.624780268Z" level=info msg="CreateContainer within sandbox \"4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 7 01:14:08.628818 containerd[1638]: time="2026-03-07T01:14:08.628735589Z" level=info msg="CreateContainer within sandbox \"66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d0dce5b72d656da9034a3bf333635a9bcad23837c0535c3629c48a52cc619d87\"" Mar 7 01:14:08.629205 containerd[1638]: time="2026-03-07T01:14:08.629180909Z" level=info msg="StartContainer for \"d0dce5b72d656da9034a3bf333635a9bcad23837c0535c3629c48a52cc619d87\"" Mar 7 01:14:08.641534 containerd[1638]: time="2026-03-07T01:14:08.641479091Z" level=info msg="CreateContainer within sandbox \"4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8c7ce796d9b893b8267b719df9099f4c849ddc308fc120e33dc87eafee5edcc5\"" Mar 7 01:14:08.642602 containerd[1638]: time="2026-03-07T01:14:08.641869971Z" level=info msg="StartContainer for \"8c7ce796d9b893b8267b719df9099f4c849ddc308fc120e33dc87eafee5edcc5\"" Mar 7 01:14:08.704220 containerd[1638]: time="2026-03-07T01:14:08.704193949Z" level=info msg="StartContainer for \"d0dce5b72d656da9034a3bf333635a9bcad23837c0535c3629c48a52cc619d87\" returns successfully" Mar 7 01:14:08.745482 containerd[1638]: time="2026-03-07T01:14:08.745271804Z" level=info msg="StartContainer for \"8c7ce796d9b893b8267b719df9099f4c849ddc308fc120e33dc87eafee5edcc5\" returns successfully" Mar 7 01:14:09.042830 containerd[1638]: time="2026-03-07T01:14:09.041802704Z" level=info msg="StopPodSandbox for \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\"" Mar 7 01:14:09.046409 containerd[1638]: time="2026-03-07T01:14:09.041811364Z" level=info msg="StopPodSandbox for \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\"" Mar 7 01:14:09.090197 systemd-networkd[1251]: cali295ecf6c436: Gained IPv6LL Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.133 [INFO][5295] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.133 [INFO][5295] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" iface="eth0" netns="/var/run/netns/cni-e6204774-a379-ee67-e1d5-26cef5bb00bb" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.138 [INFO][5295] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" iface="eth0" netns="/var/run/netns/cni-e6204774-a379-ee67-e1d5-26cef5bb00bb" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.143 [INFO][5295] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" iface="eth0" netns="/var/run/netns/cni-e6204774-a379-ee67-e1d5-26cef5bb00bb" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.143 [INFO][5295] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.143 [INFO][5295] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.171 [INFO][5312] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.171 [INFO][5312] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.171 [INFO][5312] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.178 [WARNING][5312] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.178 [INFO][5312] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.181 [INFO][5312] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:09.194398 containerd[1638]: 2026-03-07 01:14:09.190 [INFO][5295] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:09.197129 containerd[1638]: time="2026-03-07T01:14:09.196702717Z" level=info msg="TearDown network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\" successfully" Mar 7 01:14:09.197129 containerd[1638]: time="2026-03-07T01:14:09.196723637Z" level=info msg="StopPodSandbox for \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\" returns successfully" Mar 7 01:14:09.197323 containerd[1638]: time="2026-03-07T01:14:09.197309508Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-xbz8b,Uid:1fd3aff8-915e-4a01-9844-abd28ee38035,Namespace:calico-system,Attempt:1,}" Mar 7 01:14:09.199326 systemd[1]: run-netns-cni\x2de6204774\x2da379\x2dee67\x2de1d5\x2d26cef5bb00bb.mount: Deactivated successfully. Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.143 [INFO][5296] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.143 [INFO][5296] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" iface="eth0" netns="/var/run/netns/cni-49ef329e-13f2-c530-37cb-9aa5c55b4592" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.144 [INFO][5296] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" iface="eth0" netns="/var/run/netns/cni-49ef329e-13f2-c530-37cb-9aa5c55b4592" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.144 [INFO][5296] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" iface="eth0" netns="/var/run/netns/cni-49ef329e-13f2-c530-37cb-9aa5c55b4592" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.144 [INFO][5296] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.144 [INFO][5296] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.203 [INFO][5317] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.204 [INFO][5317] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.204 [INFO][5317] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.210 [WARNING][5317] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.210 [INFO][5317] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.211 [INFO][5317] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:09.226389 containerd[1638]: 2026-03-07 01:14:09.222 [INFO][5296] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:09.228879 containerd[1638]: time="2026-03-07T01:14:09.228578883Z" level=info msg="TearDown network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\" successfully" Mar 7 01:14:09.228879 containerd[1638]: time="2026-03-07T01:14:09.228598953Z" level=info msg="StopPodSandbox for \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\" returns successfully" Mar 7 01:14:09.229655 containerd[1638]: time="2026-03-07T01:14:09.229346822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-689889c887-zgnj6,Uid:571d5c28-cf64-4799-a2c5-6c5bf2b8ab06,Namespace:calico-system,Attempt:1,}" Mar 7 01:14:09.230709 systemd[1]: run-netns-cni\x2d49ef329e\x2d13f2\x2dc530\x2d37cb\x2d9aa5c55b4592.mount: Deactivated successfully. Mar 7 01:14:09.258355 kubelet[2747]: I0307 01:14:09.258338 2747 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:14:09.277827 kubelet[2747]: I0307 01:14:09.276242 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-98kpm" podStartSLOduration=39.276229959 podStartE2EDuration="39.276229959s" podCreationTimestamp="2026-03-07 01:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:14:09.264146467 +0000 UTC m=+44.331058184" watchObservedRunningTime="2026-03-07 01:14:09.276229959 +0000 UTC m=+44.343141686" Mar 7 01:14:09.277827 kubelet[2747]: I0307 01:14:09.276396 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-4xvx5" podStartSLOduration=39.276393319 podStartE2EDuration="39.276393319s" podCreationTimestamp="2026-03-07 01:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:14:09.275379289 +0000 UTC m=+44.342291006" watchObservedRunningTime="2026-03-07 01:14:09.276393319 +0000 UTC m=+44.343305046" Mar 7 01:14:09.464310 systemd-networkd[1251]: cali394d6dfe0f2: Link UP Mar 7 01:14:09.471862 systemd-networkd[1251]: cali394d6dfe0f2: Gained carrier Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.345 [INFO][5325] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0 csi-node-driver- calico-system 1fd3aff8-915e-4a01-9844-abd28ee38035 1010 0 2026-03-07 01:13:40 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6d9d697c7c k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081-3-6-n-0c8881e772 csi-node-driver-xbz8b eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali394d6dfe0f2 [] [] }} ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Namespace="calico-system" Pod="csi-node-driver-xbz8b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.345 [INFO][5325] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Namespace="calico-system" Pod="csi-node-driver-xbz8b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.420 [INFO][5348] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" HandleID="k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.429 [INFO][5348] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" HandleID="k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fea0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-0c8881e772", "pod":"csi-node-driver-xbz8b", "timestamp":"2026-03-07 01:14:09.420766771 +0000 UTC"}, Hostname:"ci-4081-3-6-n-0c8881e772", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0003a0580)} Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.429 [INFO][5348] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.429 [INFO][5348] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.429 [INFO][5348] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-0c8881e772' Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.431 [INFO][5348] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.435 [INFO][5348] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.439 [INFO][5348] ipam/ipam.go 526: Trying affinity for 192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.440 [INFO][5348] ipam/ipam.go 160: Attempting to load block cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.442 [INFO][5348] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.442 [INFO][5348] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.113.0/26 handle="k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.444 [INFO][5348] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.449 [INFO][5348] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.113.0/26 handle="k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.455 [INFO][5348] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.113.7/26] block=192.168.113.0/26 handle="k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.455 [INFO][5348] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.113.7/26] handle="k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.455 [INFO][5348] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:09.503569 containerd[1638]: 2026-03-07 01:14:09.455 [INFO][5348] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.113.7/26] IPv6=[] ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" HandleID="k8s-pod-network.03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.504044 containerd[1638]: 2026-03-07 01:14:09.458 [INFO][5325] cni-plugin/k8s.go 418: Populated endpoint ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Namespace="calico-system" Pod="csi-node-driver-xbz8b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"1fd3aff8-915e-4a01-9844-abd28ee38035", ResourceVersion:"1010", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"", Pod:"csi-node-driver-xbz8b", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.113.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali394d6dfe0f2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:09.504044 containerd[1638]: 2026-03-07 01:14:09.458 [INFO][5325] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.113.7/32] ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Namespace="calico-system" Pod="csi-node-driver-xbz8b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.504044 containerd[1638]: 2026-03-07 01:14:09.458 [INFO][5325] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali394d6dfe0f2 ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Namespace="calico-system" Pod="csi-node-driver-xbz8b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.504044 containerd[1638]: 2026-03-07 01:14:09.460 [INFO][5325] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Namespace="calico-system" Pod="csi-node-driver-xbz8b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.504044 containerd[1638]: 2026-03-07 01:14:09.460 [INFO][5325] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Namespace="calico-system" Pod="csi-node-driver-xbz8b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"1fd3aff8-915e-4a01-9844-abd28ee38035", ResourceVersion:"1010", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e", Pod:"csi-node-driver-xbz8b", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.113.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali394d6dfe0f2", MAC:"92:ac:c8:76:3c:96", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:09.504044 containerd[1638]: 2026-03-07 01:14:09.468 [INFO][5325] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e" Namespace="calico-system" Pod="csi-node-driver-xbz8b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:09.585598 containerd[1638]: time="2026-03-07T01:14:09.585091305Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:14:09.585598 containerd[1638]: time="2026-03-07T01:14:09.585134745Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:14:09.585598 containerd[1638]: time="2026-03-07T01:14:09.585146535Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:09.585598 containerd[1638]: time="2026-03-07T01:14:09.585225535Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:09.594285 systemd-networkd[1251]: cali1b60106aa2d: Link UP Mar 7 01:14:09.594437 systemd-networkd[1251]: cali1b60106aa2d: Gained carrier Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.366 [INFO][5334] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0 calico-apiserver-689889c887- calico-system 571d5c28-cf64-4799-a2c5-6c5bf2b8ab06 1011 0 2026-03-07 01:13:39 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:689889c887 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081-3-6-n-0c8881e772 calico-apiserver-689889c887-zgnj6 eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali1b60106aa2d [] [] }} ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Namespace="calico-system" Pod="calico-apiserver-689889c887-zgnj6" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.367 [INFO][5334] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Namespace="calico-system" Pod="calico-apiserver-689889c887-zgnj6" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.442 [INFO][5355] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" HandleID="k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.450 [INFO][5355] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" HandleID="k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000305ea0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-6-n-0c8881e772", "pod":"calico-apiserver-689889c887-zgnj6", "timestamp":"2026-03-07 01:14:09.442334735 +0000 UTC"}, Hostname:"ci-4081-3-6-n-0c8881e772", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0002c1760)} Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.451 [INFO][5355] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.456 [INFO][5355] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.456 [INFO][5355] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-6-n-0c8881e772' Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.534 [INFO][5355] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.541 [INFO][5355] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.546 [INFO][5355] ipam/ipam.go 526: Trying affinity for 192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.548 [INFO][5355] ipam/ipam.go 160: Attempting to load block cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.549 [INFO][5355] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.113.0/26 host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.549 [INFO][5355] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.113.0/26 handle="k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.554 [INFO][5355] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.564 [INFO][5355] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.113.0/26 handle="k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.575 [INFO][5355] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.113.8/26] block=192.168.113.0/26 handle="k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.575 [INFO][5355] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.113.8/26] handle="k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" host="ci-4081-3-6-n-0c8881e772" Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.576 [INFO][5355] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:09.644310 containerd[1638]: 2026-03-07 01:14:09.576 [INFO][5355] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.113.8/26] IPv6=[] ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" HandleID="k8s-pod-network.7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.644763 containerd[1638]: 2026-03-07 01:14:09.587 [INFO][5334] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Namespace="calico-system" Pod="calico-apiserver-689889c887-zgnj6" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0", GenerateName:"calico-apiserver-689889c887-", Namespace:"calico-system", SelfLink:"", UID:"571d5c28-cf64-4799-a2c5-6c5bf2b8ab06", ResourceVersion:"1011", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"689889c887", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"", Pod:"calico-apiserver-689889c887-zgnj6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.113.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali1b60106aa2d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:09.644763 containerd[1638]: 2026-03-07 01:14:09.588 [INFO][5334] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.113.8/32] ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Namespace="calico-system" Pod="calico-apiserver-689889c887-zgnj6" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.644763 containerd[1638]: 2026-03-07 01:14:09.588 [INFO][5334] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1b60106aa2d ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Namespace="calico-system" Pod="calico-apiserver-689889c887-zgnj6" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.644763 containerd[1638]: 2026-03-07 01:14:09.602 [INFO][5334] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Namespace="calico-system" Pod="calico-apiserver-689889c887-zgnj6" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.644763 containerd[1638]: 2026-03-07 01:14:09.604 [INFO][5334] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Namespace="calico-system" Pod="calico-apiserver-689889c887-zgnj6" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0", GenerateName:"calico-apiserver-689889c887-", Namespace:"calico-system", SelfLink:"", UID:"571d5c28-cf64-4799-a2c5-6c5bf2b8ab06", ResourceVersion:"1011", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"689889c887", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f", Pod:"calico-apiserver-689889c887-zgnj6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.113.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali1b60106aa2d", MAC:"2a:b7:13:b4:06:ed", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:09.644763 containerd[1638]: 2026-03-07 01:14:09.620 [INFO][5334] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f" Namespace="calico-system" Pod="calico-apiserver-689889c887-zgnj6" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:09.706158 containerd[1638]: time="2026-03-07T01:14:09.706130583Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-xbz8b,Uid:1fd3aff8-915e-4a01-9844-abd28ee38035,Namespace:calico-system,Attempt:1,} returns sandbox id \"03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e\"" Mar 7 01:14:09.708159 containerd[1638]: time="2026-03-07T01:14:09.707976034Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:14:09.708159 containerd[1638]: time="2026-03-07T01:14:09.708023954Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:14:09.708159 containerd[1638]: time="2026-03-07T01:14:09.708038554Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:09.708159 containerd[1638]: time="2026-03-07T01:14:09.708136254Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:14:09.782186 containerd[1638]: time="2026-03-07T01:14:09.782110785Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-689889c887-zgnj6,Uid:571d5c28-cf64-4799-a2c5-6c5bf2b8ab06,Namespace:calico-system,Attempt:1,} returns sandbox id \"7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f\"" Mar 7 01:14:09.786189 containerd[1638]: time="2026-03-07T01:14:09.786077106Z" level=info msg="CreateContainer within sandbox \"7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 7 01:14:09.803520 containerd[1638]: time="2026-03-07T01:14:09.803481108Z" level=info msg="CreateContainer within sandbox \"7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"f4f620d9f3fff8a9f7ec5304786eced6ee15d261c102bc801dd96ada7e8ad6fe\"" Mar 7 01:14:09.804580 containerd[1638]: time="2026-03-07T01:14:09.804563698Z" level=info msg="StartContainer for \"f4f620d9f3fff8a9f7ec5304786eced6ee15d261c102bc801dd96ada7e8ad6fe\"" Mar 7 01:14:09.857719 systemd-networkd[1251]: calib83eb51f121: Gained IPv6LL Mar 7 01:14:09.902196 containerd[1638]: time="2026-03-07T01:14:09.902165323Z" level=info msg="StartContainer for \"f4f620d9f3fff8a9f7ec5304786eced6ee15d261c102bc801dd96ada7e8ad6fe\" returns successfully" Mar 7 01:14:09.921947 systemd-networkd[1251]: calif37b16151c8: Gained IPv6LL Mar 7 01:14:10.193164 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1238861656.mount: Deactivated successfully. Mar 7 01:14:10.278752 kubelet[2747]: I0307 01:14:10.278177 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-689889c887-zgnj6" podStartSLOduration=31.278156544 podStartE2EDuration="31.278156544s" podCreationTimestamp="2026-03-07 01:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:14:10.276279604 +0000 UTC m=+45.343191331" watchObservedRunningTime="2026-03-07 01:14:10.278156544 +0000 UTC m=+45.345068261" Mar 7 01:14:10.626360 systemd-networkd[1251]: cali394d6dfe0f2: Gained IPv6LL Mar 7 01:14:10.645103 containerd[1638]: time="2026-03-07T01:14:10.644556705Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:10.646208 containerd[1638]: time="2026-03-07T01:14:10.646159685Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.4: active requests=0, bytes read=55623386" Mar 7 01:14:10.647153 containerd[1638]: time="2026-03-07T01:14:10.646913915Z" level=info msg="ImageCreate event name:\"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:10.652175 containerd[1638]: time="2026-03-07T01:14:10.652147586Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:10.653761 containerd[1638]: time="2026-03-07T01:14:10.653437086Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" with image id \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\", size \"55623232\" in 2.906773199s" Mar 7 01:14:10.653761 containerd[1638]: time="2026-03-07T01:14:10.653462646Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" returns image reference \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\"" Mar 7 01:14:10.657889 containerd[1638]: time="2026-03-07T01:14:10.657868497Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\"" Mar 7 01:14:10.661584 containerd[1638]: time="2026-03-07T01:14:10.661265297Z" level=info msg="CreateContainer within sandbox \"1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Mar 7 01:14:10.685584 containerd[1638]: time="2026-03-07T01:14:10.685062971Z" level=info msg="CreateContainer within sandbox \"1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"7c7a8e38c05d4d3d85edb57ad8508f95b57b2e29d482329fc3b28162af264bc0\"" Mar 7 01:14:10.685913 containerd[1638]: time="2026-03-07T01:14:10.685899052Z" level=info msg="StartContainer for \"7c7a8e38c05d4d3d85edb57ad8508f95b57b2e29d482329fc3b28162af264bc0\"" Mar 7 01:14:10.764760 containerd[1638]: time="2026-03-07T01:14:10.764594255Z" level=info msg="StartContainer for \"7c7a8e38c05d4d3d85edb57ad8508f95b57b2e29d482329fc3b28162af264bc0\" returns successfully" Mar 7 01:14:10.817802 systemd-networkd[1251]: cali1b60106aa2d: Gained IPv6LL Mar 7 01:14:11.280116 kubelet[2747]: I0307 01:14:11.279450 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-5b85766d88-9qlr7" podStartSLOduration=27.08136481 podStartE2EDuration="32.279433235s" podCreationTimestamp="2026-03-07 01:13:39 +0000 UTC" firstStartedPulling="2026-03-07 01:14:05.458117642 +0000 UTC m=+40.525029369" lastFinishedPulling="2026-03-07 01:14:10.656186067 +0000 UTC m=+45.723097794" observedRunningTime="2026-03-07 01:14:11.279197875 +0000 UTC m=+46.346109602" watchObservedRunningTime="2026-03-07 01:14:11.279433235 +0000 UTC m=+46.346344962" Mar 7 01:14:11.462803 systemd-journald[1177]: Under memory pressure, flushing caches. Mar 7 01:14:11.457884 systemd-resolved[1515]: Under memory pressure, flushing caches. Mar 7 01:14:11.457957 systemd-resolved[1515]: Flushed all caches. Mar 7 01:14:12.283573 kubelet[2747]: I0307 01:14:12.281118 2747 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:14:12.788189 containerd[1638]: time="2026-03-07T01:14:12.788148051Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:12.789327 containerd[1638]: time="2026-03-07T01:14:12.789176001Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.4: active requests=0, bytes read=52406348" Mar 7 01:14:12.790252 containerd[1638]: time="2026-03-07T01:14:12.790216432Z" level=info msg="ImageCreate event name:\"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:12.792498 containerd[1638]: time="2026-03-07T01:14:12.792447022Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:12.793317 containerd[1638]: time="2026-03-07T01:14:12.792967882Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" with image id \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\", size \"53962361\" in 2.134327655s" Mar 7 01:14:12.793317 containerd[1638]: time="2026-03-07T01:14:12.792991692Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" returns image reference \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\"" Mar 7 01:14:12.795236 containerd[1638]: time="2026-03-07T01:14:12.795223422Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\"" Mar 7 01:14:12.808407 containerd[1638]: time="2026-03-07T01:14:12.808372915Z" level=info msg="CreateContainer within sandbox \"40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Mar 7 01:14:12.818369 containerd[1638]: time="2026-03-07T01:14:12.818300977Z" level=info msg="CreateContainer within sandbox \"40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"ca2bb687d00bffd48b4dd3b097c611544b1766fe753a5c050a4078a0170f3de3\"" Mar 7 01:14:12.820717 containerd[1638]: time="2026-03-07T01:14:12.819690227Z" level=info msg="StartContainer for \"ca2bb687d00bffd48b4dd3b097c611544b1766fe753a5c050a4078a0170f3de3\"" Mar 7 01:14:12.883730 containerd[1638]: time="2026-03-07T01:14:12.883687200Z" level=info msg="StartContainer for \"ca2bb687d00bffd48b4dd3b097c611544b1766fe753a5c050a4078a0170f3de3\" returns successfully" Mar 7 01:14:13.304874 kubelet[2747]: I0307 01:14:13.304802 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5dc64846bc-k762g" podStartSLOduration=27.959407478 podStartE2EDuration="33.304778027s" podCreationTimestamp="2026-03-07 01:13:40 +0000 UTC" firstStartedPulling="2026-03-07 01:14:07.448346043 +0000 UTC m=+42.515257760" lastFinishedPulling="2026-03-07 01:14:12.793716592 +0000 UTC m=+47.860628309" observedRunningTime="2026-03-07 01:14:13.303707157 +0000 UTC m=+48.370618914" watchObservedRunningTime="2026-03-07 01:14:13.304778027 +0000 UTC m=+48.371689794" Mar 7 01:14:14.447504 containerd[1638]: time="2026-03-07T01:14:14.447244382Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:14.449275 containerd[1638]: time="2026-03-07T01:14:14.449224343Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.4: active requests=0, bytes read=8792502" Mar 7 01:14:14.450532 containerd[1638]: time="2026-03-07T01:14:14.450495312Z" level=info msg="ImageCreate event name:\"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:14.456083 containerd[1638]: time="2026-03-07T01:14:14.455564554Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:14.457959 containerd[1638]: time="2026-03-07T01:14:14.457933855Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.4\" with image id \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\", size \"10348547\" in 1.662608703s" Mar 7 01:14:14.458050 containerd[1638]: time="2026-03-07T01:14:14.458038155Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\" returns image reference \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\"" Mar 7 01:14:14.462943 containerd[1638]: time="2026-03-07T01:14:14.462906046Z" level=info msg="CreateContainer within sandbox \"03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 7 01:14:14.479751 containerd[1638]: time="2026-03-07T01:14:14.479716670Z" level=info msg="CreateContainer within sandbox \"03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"01d149cc56fdacb8c097fe039673980b8fbd9556c1c152bb83ce8fb915c460f0\"" Mar 7 01:14:14.480181 containerd[1638]: time="2026-03-07T01:14:14.480067320Z" level=info msg="StartContainer for \"01d149cc56fdacb8c097fe039673980b8fbd9556c1c152bb83ce8fb915c460f0\"" Mar 7 01:14:14.536028 containerd[1638]: time="2026-03-07T01:14:14.535980672Z" level=info msg="StartContainer for \"01d149cc56fdacb8c097fe039673980b8fbd9556c1c152bb83ce8fb915c460f0\" returns successfully" Mar 7 01:14:14.538914 containerd[1638]: time="2026-03-07T01:14:14.538878372Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\"" Mar 7 01:14:16.267016 containerd[1638]: time="2026-03-07T01:14:16.266918378Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:16.268130 containerd[1638]: time="2026-03-07T01:14:16.267976398Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4: active requests=0, bytes read=14704317" Mar 7 01:14:16.268982 containerd[1638]: time="2026-03-07T01:14:16.268906469Z" level=info msg="ImageCreate event name:\"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:16.270603 containerd[1638]: time="2026-03-07T01:14:16.270573539Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:14:16.271055 containerd[1638]: time="2026-03-07T01:14:16.271031859Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" with image id \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\", size \"16260314\" in 1.732103877s" Mar 7 01:14:16.271096 containerd[1638]: time="2026-03-07T01:14:16.271057189Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" returns image reference \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\"" Mar 7 01:14:16.274626 containerd[1638]: time="2026-03-07T01:14:16.274528750Z" level=info msg="CreateContainer within sandbox \"03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 7 01:14:16.289479 containerd[1638]: time="2026-03-07T01:14:16.289449253Z" level=info msg="CreateContainer within sandbox \"03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"906ab2a8a027482cfc0f766bd48e55279d957abc557821ef68548dc05520b3c0\"" Mar 7 01:14:16.290884 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3562886304.mount: Deactivated successfully. Mar 7 01:14:16.292264 containerd[1638]: time="2026-03-07T01:14:16.292243024Z" level=info msg="StartContainer for \"906ab2a8a027482cfc0f766bd48e55279d957abc557821ef68548dc05520b3c0\"" Mar 7 01:14:16.342258 containerd[1638]: time="2026-03-07T01:14:16.342220866Z" level=info msg="StartContainer for \"906ab2a8a027482cfc0f766bd48e55279d957abc557821ef68548dc05520b3c0\" returns successfully" Mar 7 01:14:17.117740 kubelet[2747]: I0307 01:14:17.117536 2747 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 7 01:14:17.119713 kubelet[2747]: I0307 01:14:17.119679 2747 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 7 01:14:17.317003 kubelet[2747]: I0307 01:14:17.316881 2747 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-xbz8b" podStartSLOduration=30.754475886 podStartE2EDuration="37.316863291s" podCreationTimestamp="2026-03-07 01:13:40 +0000 UTC" firstStartedPulling="2026-03-07 01:14:09.709328594 +0000 UTC m=+44.776240311" lastFinishedPulling="2026-03-07 01:14:16.271715989 +0000 UTC m=+51.338627716" observedRunningTime="2026-03-07 01:14:17.316203041 +0000 UTC m=+52.383114808" watchObservedRunningTime="2026-03-07 01:14:17.316863291 +0000 UTC m=+52.383775048" Mar 7 01:14:18.960173 kubelet[2747]: I0307 01:14:18.959824 2747 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:14:19.043864 systemd[1]: run-containerd-runc-k8s.io-7c7a8e38c05d4d3d85edb57ad8508f95b57b2e29d482329fc3b28162af264bc0-runc.ctolln.mount: Deactivated successfully. Mar 7 01:14:25.005470 containerd[1638]: time="2026-03-07T01:14:25.005374973Z" level=info msg="StopPodSandbox for \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\"" Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.046 [WARNING][5836] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0", GenerateName:"calico-apiserver-689889c887-", Namespace:"calico-system", SelfLink:"", UID:"a9f53e47-46dd-420a-9df1-28f9cae95a81", ResourceVersion:"992", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"689889c887", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587", Pod:"calico-apiserver-689889c887-ntpwf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.113.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali42e2d916126", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.046 [INFO][5836] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.046 [INFO][5836] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" iface="eth0" netns="" Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.046 [INFO][5836] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.046 [INFO][5836] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.075 [INFO][5844] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.075 [INFO][5844] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.075 [INFO][5844] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.079 [WARNING][5844] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.079 [INFO][5844] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.081 [INFO][5844] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.086720 containerd[1638]: 2026-03-07 01:14:25.084 [INFO][5836] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:25.087066 containerd[1638]: time="2026-03-07T01:14:25.086743367Z" level=info msg="TearDown network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\" successfully" Mar 7 01:14:25.087066 containerd[1638]: time="2026-03-07T01:14:25.086761456Z" level=info msg="StopPodSandbox for \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\" returns successfully" Mar 7 01:14:25.087333 containerd[1638]: time="2026-03-07T01:14:25.087309439Z" level=info msg="RemovePodSandbox for \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\"" Mar 7 01:14:25.087333 containerd[1638]: time="2026-03-07T01:14:25.087332509Z" level=info msg="Forcibly stopping sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\"" Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.111 [WARNING][5858] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0", GenerateName:"calico-apiserver-689889c887-", Namespace:"calico-system", SelfLink:"", UID:"a9f53e47-46dd-420a-9df1-28f9cae95a81", ResourceVersion:"992", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"689889c887", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"4835911f81599fa7d14a8da38dfa4d05e6cfa7a7762964b4791839a141b9f587", Pod:"calico-apiserver-689889c887-ntpwf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.113.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali42e2d916126", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.112 [INFO][5858] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.112 [INFO][5858] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" iface="eth0" netns="" Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.112 [INFO][5858] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.112 [INFO][5858] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.127 [INFO][5865] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.127 [INFO][5865] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.127 [INFO][5865] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.133 [WARNING][5865] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.133 [INFO][5865] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" HandleID="k8s-pod-network.e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--ntpwf-eth0" Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.134 [INFO][5865] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.138388 containerd[1638]: 2026-03-07 01:14:25.136 [INFO][5858] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd" Mar 7 01:14:25.138776 containerd[1638]: time="2026-03-07T01:14:25.138417163Z" level=info msg="TearDown network for sandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\" successfully" Mar 7 01:14:25.142998 containerd[1638]: time="2026-03-07T01:14:25.142964136Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:14:25.143068 containerd[1638]: time="2026-03-07T01:14:25.143013294Z" level=info msg="RemovePodSandbox \"e77647c3a5bb1682200ff934b3c4b4126c0702638cd53ed0fe1bd1beba2297bd\" returns successfully" Mar 7 01:14:25.143640 containerd[1638]: time="2026-03-07T01:14:25.143418331Z" level=info msg="StopPodSandbox for \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\"" Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.168 [WARNING][5881] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"1d72866e-1f81-4c34-a51f-96e1e7fc603a", ResourceVersion:"1020", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7", Pod:"coredns-674b8bbfcf-98kpm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.113.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif37b16151c8", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.169 [INFO][5881] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.169 [INFO][5881] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" iface="eth0" netns="" Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.169 [INFO][5881] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.169 [INFO][5881] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.184 [INFO][5888] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.185 [INFO][5888] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.185 [INFO][5888] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.190 [WARNING][5888] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.190 [INFO][5888] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.191 [INFO][5888] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.195418 containerd[1638]: 2026-03-07 01:14:25.193 [INFO][5881] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:25.196032 containerd[1638]: time="2026-03-07T01:14:25.195639279Z" level=info msg="TearDown network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\" successfully" Mar 7 01:14:25.196032 containerd[1638]: time="2026-03-07T01:14:25.195662749Z" level=info msg="StopPodSandbox for \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\" returns successfully" Mar 7 01:14:25.196485 containerd[1638]: time="2026-03-07T01:14:25.196210001Z" level=info msg="RemovePodSandbox for \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\"" Mar 7 01:14:25.196485 containerd[1638]: time="2026-03-07T01:14:25.196236590Z" level=info msg="Forcibly stopping sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\"" Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.231 [WARNING][5903] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"1d72866e-1f81-4c34-a51f-96e1e7fc603a", ResourceVersion:"1020", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"4f9cb7215975b4dc506e9e8f72130494c5c911a86ce276b4c853545e788e62d7", Pod:"coredns-674b8bbfcf-98kpm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.113.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif37b16151c8", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.231 [INFO][5903] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.231 [INFO][5903] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" iface="eth0" netns="" Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.231 [INFO][5903] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.231 [INFO][5903] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.255 [INFO][5926] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.255 [INFO][5926] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.255 [INFO][5926] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.260 [WARNING][5926] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.260 [INFO][5926] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" HandleID="k8s-pod-network.3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--98kpm-eth0" Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.261 [INFO][5926] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.269100 containerd[1638]: 2026-03-07 01:14:25.264 [INFO][5903] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb" Mar 7 01:14:25.270267 containerd[1638]: time="2026-03-07T01:14:25.269604892Z" level=info msg="TearDown network for sandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\" successfully" Mar 7 01:14:25.273633 containerd[1638]: time="2026-03-07T01:14:25.273605933Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:14:25.273694 containerd[1638]: time="2026-03-07T01:14:25.273656451Z" level=info msg="RemovePodSandbox \"3c669c0a0d54f4ac164318f0e0a0c3669fe6489476059af53e5a57a752bda0bb\" returns successfully" Mar 7 01:14:25.274177 containerd[1638]: time="2026-03-07T01:14:25.273951231Z" level=info msg="StopPodSandbox for \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\"" Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.298 [WARNING][5944] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"ca853831-8f8c-4ad5-a296-b442696cf9d1", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042", Pod:"goldmane-5b85766d88-9qlr7", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.113.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calidb8eca2181a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.298 [INFO][5944] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.298 [INFO][5944] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" iface="eth0" netns="" Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.298 [INFO][5944] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.298 [INFO][5944] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.313 [INFO][5952] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.313 [INFO][5952] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.313 [INFO][5952] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.317 [WARNING][5952] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.317 [INFO][5952] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.318 [INFO][5952] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.322740 containerd[1638]: 2026-03-07 01:14:25.320 [INFO][5944] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:25.323448 containerd[1638]: time="2026-03-07T01:14:25.322765370Z" level=info msg="TearDown network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\" successfully" Mar 7 01:14:25.323448 containerd[1638]: time="2026-03-07T01:14:25.322781690Z" level=info msg="StopPodSandbox for \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\" returns successfully" Mar 7 01:14:25.323448 containerd[1638]: time="2026-03-07T01:14:25.322954074Z" level=info msg="RemovePodSandbox for \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\"" Mar 7 01:14:25.323448 containerd[1638]: time="2026-03-07T01:14:25.322982303Z" level=info msg="Forcibly stopping sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\"" Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.344 [WARNING][5967] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"ca853831-8f8c-4ad5-a296-b442696cf9d1", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"1b842d24c8041edecc6a07309b34af8cfc808736efa34b6697187751b5cc0042", Pod:"goldmane-5b85766d88-9qlr7", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.113.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calidb8eca2181a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.344 [INFO][5967] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.344 [INFO][5967] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" iface="eth0" netns="" Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.344 [INFO][5967] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.344 [INFO][5967] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.362 [INFO][5974] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.362 [INFO][5974] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.362 [INFO][5974] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.366 [WARNING][5974] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.366 [INFO][5974] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" HandleID="k8s-pod-network.ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Workload="ci--4081--3--6--n--0c8881e772-k8s-goldmane--5b85766d88--9qlr7-eth0" Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.367 [INFO][5974] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.370452 containerd[1638]: 2026-03-07 01:14:25.368 [INFO][5967] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc" Mar 7 01:14:25.370805 containerd[1638]: time="2026-03-07T01:14:25.370501143Z" level=info msg="TearDown network for sandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\" successfully" Mar 7 01:14:25.374582 containerd[1638]: time="2026-03-07T01:14:25.374511363Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:14:25.374798 containerd[1638]: time="2026-03-07T01:14:25.374649059Z" level=info msg="RemovePodSandbox \"ce569449164898e5a3dad9806f1e5e840c8544f4c90355c3d2d0f80e025b5ccc\" returns successfully" Mar 7 01:14:25.375293 containerd[1638]: time="2026-03-07T01:14:25.375093444Z" level=info msg="StopPodSandbox for \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\"" Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.411 [WARNING][5988] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0", GenerateName:"calico-apiserver-689889c887-", Namespace:"calico-system", SelfLink:"", UID:"571d5c28-cf64-4799-a2c5-6c5bf2b8ab06", ResourceVersion:"1044", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"689889c887", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f", Pod:"calico-apiserver-689889c887-zgnj6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.113.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali1b60106aa2d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.411 [INFO][5988] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.411 [INFO][5988] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" iface="eth0" netns="" Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.411 [INFO][5988] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.411 [INFO][5988] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.428 [INFO][5996] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.429 [INFO][5996] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.429 [INFO][5996] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.433 [WARNING][5996] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.433 [INFO][5996] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.434 [INFO][5996] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.437489 containerd[1638]: 2026-03-07 01:14:25.435 [INFO][5988] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:25.437865 containerd[1638]: time="2026-03-07T01:14:25.437518002Z" level=info msg="TearDown network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\" successfully" Mar 7 01:14:25.437865 containerd[1638]: time="2026-03-07T01:14:25.437574940Z" level=info msg="StopPodSandbox for \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\" returns successfully" Mar 7 01:14:25.437916 containerd[1638]: time="2026-03-07T01:14:25.437896920Z" level=info msg="RemovePodSandbox for \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\"" Mar 7 01:14:25.437993 containerd[1638]: time="2026-03-07T01:14:25.437913880Z" level=info msg="Forcibly stopping sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\"" Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.461 [WARNING][6010] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0", GenerateName:"calico-apiserver-689889c887-", Namespace:"calico-system", SelfLink:"", UID:"571d5c28-cf64-4799-a2c5-6c5bf2b8ab06", ResourceVersion:"1044", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"689889c887", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"7539cd21ccdefdca6bf659334802aff53c628eff8569f4e9d584220b27d9fa7f", Pod:"calico-apiserver-689889c887-zgnj6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.113.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali1b60106aa2d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.461 [INFO][6010] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.461 [INFO][6010] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" iface="eth0" netns="" Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.461 [INFO][6010] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.461 [INFO][6010] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.475 [INFO][6017] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.476 [INFO][6017] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.476 [INFO][6017] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.480 [WARNING][6017] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.480 [INFO][6017] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" HandleID="k8s-pod-network.37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--apiserver--689889c887--zgnj6-eth0" Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.481 [INFO][6017] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.484386 containerd[1638]: 2026-03-07 01:14:25.482 [INFO][6010] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971" Mar 7 01:14:25.484846 containerd[1638]: time="2026-03-07T01:14:25.484391973Z" level=info msg="TearDown network for sandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\" successfully" Mar 7 01:14:25.488041 containerd[1638]: time="2026-03-07T01:14:25.488015365Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:14:25.488088 containerd[1638]: time="2026-03-07T01:14:25.488061614Z" level=info msg="RemovePodSandbox \"37f107cb5e8ea3fc60292cc71ce18cd936afa878e5e468d70e242ee00a923971\" returns successfully" Mar 7 01:14:25.488399 containerd[1638]: time="2026-03-07T01:14:25.488384444Z" level=info msg="StopPodSandbox for \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\"" Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.513 [WARNING][6031] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"99f62e43-c05f-4b23-90d0-600585809c83", ResourceVersion:"1016", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b", Pod:"coredns-674b8bbfcf-4xvx5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.113.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calib83eb51f121", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.513 [INFO][6031] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.513 [INFO][6031] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" iface="eth0" netns="" Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.514 [INFO][6031] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.514 [INFO][6031] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.528 [INFO][6038] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.528 [INFO][6038] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.528 [INFO][6038] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.532 [WARNING][6038] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.532 [INFO][6038] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.533 [INFO][6038] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.537040 containerd[1638]: 2026-03-07 01:14:25.535 [INFO][6031] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:25.537040 containerd[1638]: time="2026-03-07T01:14:25.537016088Z" level=info msg="TearDown network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\" successfully" Mar 7 01:14:25.537040 containerd[1638]: time="2026-03-07T01:14:25.537037218Z" level=info msg="StopPodSandbox for \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\" returns successfully" Mar 7 01:14:25.538195 containerd[1638]: time="2026-03-07T01:14:25.538159541Z" level=info msg="RemovePodSandbox for \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\"" Mar 7 01:14:25.538195 containerd[1638]: time="2026-03-07T01:14:25.538186690Z" level=info msg="Forcibly stopping sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\"" Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.565 [WARNING][6052] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"99f62e43-c05f-4b23-90d0-600585809c83", ResourceVersion:"1016", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"66cac48aa976d43947c479a5a1b276f67b4928d14c2bb7ed2350ea450a90a54b", Pod:"coredns-674b8bbfcf-4xvx5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.113.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calib83eb51f121", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.565 [INFO][6052] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.565 [INFO][6052] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" iface="eth0" netns="" Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.565 [INFO][6052] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.565 [INFO][6052] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.580 [INFO][6060] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.580 [INFO][6060] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.580 [INFO][6060] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.584 [WARNING][6060] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.584 [INFO][6060] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" HandleID="k8s-pod-network.e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Workload="ci--4081--3--6--n--0c8881e772-k8s-coredns--674b8bbfcf--4xvx5-eth0" Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.585 [INFO][6060] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.588831 containerd[1638]: 2026-03-07 01:14:25.587 [INFO][6052] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339" Mar 7 01:14:25.589172 containerd[1638]: time="2026-03-07T01:14:25.588858838Z" level=info msg="TearDown network for sandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\" successfully" Mar 7 01:14:25.592369 containerd[1638]: time="2026-03-07T01:14:25.592346086Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:14:25.592430 containerd[1638]: time="2026-03-07T01:14:25.592386834Z" level=info msg="RemovePodSandbox \"e2cb5079f38b39820347d9ab570afeca672859282d1e2d8d697749900c86b339\" returns successfully" Mar 7 01:14:25.592876 containerd[1638]: time="2026-03-07T01:14:25.592802820Z" level=info msg="StopPodSandbox for \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\"" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.616 [WARNING][6074] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.616 [INFO][6074] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.616 [INFO][6074] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" iface="eth0" netns="" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.616 [INFO][6074] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.616 [INFO][6074] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.632 [INFO][6082] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.632 [INFO][6082] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.632 [INFO][6082] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.637 [WARNING][6082] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.637 [INFO][6082] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.638 [INFO][6082] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.641378 containerd[1638]: 2026-03-07 01:14:25.639 [INFO][6074] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:14:25.641709 containerd[1638]: time="2026-03-07T01:14:25.641407465Z" level=info msg="TearDown network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\" successfully" Mar 7 01:14:25.641709 containerd[1638]: time="2026-03-07T01:14:25.641428775Z" level=info msg="StopPodSandbox for \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\" returns successfully" Mar 7 01:14:25.642260 containerd[1638]: time="2026-03-07T01:14:25.642026706Z" level=info msg="RemovePodSandbox for \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\"" Mar 7 01:14:25.642260 containerd[1638]: time="2026-03-07T01:14:25.642047375Z" level=info msg="Forcibly stopping sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\"" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.664 [WARNING][6097] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" WorkloadEndpoint="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.664 [INFO][6097] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.664 [INFO][6097] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" iface="eth0" netns="" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.665 [INFO][6097] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.665 [INFO][6097] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.679 [INFO][6104] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.679 [INFO][6104] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.679 [INFO][6104] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.684 [WARNING][6104] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.684 [INFO][6104] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" HandleID="k8s-pod-network.af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Workload="ci--4081--3--6--n--0c8881e772-k8s-whisker--687768496d--8mqt2-eth0" Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.685 [INFO][6104] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.688909 containerd[1638]: 2026-03-07 01:14:25.687 [INFO][6097] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b" Mar 7 01:14:25.688909 containerd[1638]: time="2026-03-07T01:14:25.688902027Z" level=info msg="TearDown network for sandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\" successfully" Mar 7 01:14:25.692096 containerd[1638]: time="2026-03-07T01:14:25.691988746Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:14:25.692096 containerd[1638]: time="2026-03-07T01:14:25.692039705Z" level=info msg="RemovePodSandbox \"af9c1b52fa39509a6e8bb766675b64137e2f6bde1fcbaa51eb123727da3bd68b\" returns successfully" Mar 7 01:14:25.692498 containerd[1638]: time="2026-03-07T01:14:25.692480061Z" level=info msg="StopPodSandbox for \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\"" Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.716 [WARNING][6119] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0", GenerateName:"calico-kube-controllers-5dc64846bc-", Namespace:"calico-system", SelfLink:"", UID:"7f399497-ed30-4f43-b4f2-62919a08223a", ResourceVersion:"1067", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5dc64846bc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa", Pod:"calico-kube-controllers-5dc64846bc-k762g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.113.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali295ecf6c436", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.716 [INFO][6119] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.716 [INFO][6119] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" iface="eth0" netns="" Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.716 [INFO][6119] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.716 [INFO][6119] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.731 [INFO][6127] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.731 [INFO][6127] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.731 [INFO][6127] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.735 [WARNING][6127] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.735 [INFO][6127] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.736 [INFO][6127] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.739439 containerd[1638]: 2026-03-07 01:14:25.737 [INFO][6119] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:25.739854 containerd[1638]: time="2026-03-07T01:14:25.739807327Z" level=info msg="TearDown network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\" successfully" Mar 7 01:14:25.739854 containerd[1638]: time="2026-03-07T01:14:25.739834036Z" level=info msg="StopPodSandbox for \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\" returns successfully" Mar 7 01:14:25.740268 containerd[1638]: time="2026-03-07T01:14:25.740245433Z" level=info msg="RemovePodSandbox for \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\"" Mar 7 01:14:25.740332 containerd[1638]: time="2026-03-07T01:14:25.740269612Z" level=info msg="Forcibly stopping sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\"" Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.764 [WARNING][6141] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0", GenerateName:"calico-kube-controllers-5dc64846bc-", Namespace:"calico-system", SelfLink:"", UID:"7f399497-ed30-4f43-b4f2-62919a08223a", ResourceVersion:"1067", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5dc64846bc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"40993cc51ff328f9e8005c970c165f3a48866dd19c5799da5a2d29fb292005aa", Pod:"calico-kube-controllers-5dc64846bc-k762g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.113.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali295ecf6c436", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.764 [INFO][6141] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.764 [INFO][6141] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" iface="eth0" netns="" Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.764 [INFO][6141] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.764 [INFO][6141] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.778 [INFO][6149] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.778 [INFO][6149] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.778 [INFO][6149] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.782 [WARNING][6149] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.782 [INFO][6149] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" HandleID="k8s-pod-network.dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Workload="ci--4081--3--6--n--0c8881e772-k8s-calico--kube--controllers--5dc64846bc--k762g-eth0" Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.783 [INFO][6149] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.786846 containerd[1638]: 2026-03-07 01:14:25.785 [INFO][6141] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e" Mar 7 01:14:25.787198 containerd[1638]: time="2026-03-07T01:14:25.786879842Z" level=info msg="TearDown network for sandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\" successfully" Mar 7 01:14:25.790666 containerd[1638]: time="2026-03-07T01:14:25.790644390Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:14:25.790720 containerd[1638]: time="2026-03-07T01:14:25.790690088Z" level=info msg="RemovePodSandbox \"dc7570940ea154a7efe4f3a62edd26fff694b56ab766405d5590a40b5117ed1e\" returns successfully" Mar 7 01:14:25.791063 containerd[1638]: time="2026-03-07T01:14:25.791042878Z" level=info msg="StopPodSandbox for \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\"" Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.814 [WARNING][6163] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"1fd3aff8-915e-4a01-9844-abd28ee38035", ResourceVersion:"1088", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e", Pod:"csi-node-driver-xbz8b", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.113.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali394d6dfe0f2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.814 [INFO][6163] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.814 [INFO][6163] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" iface="eth0" netns="" Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.814 [INFO][6163] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.814 [INFO][6163] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.830 [INFO][6171] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.830 [INFO][6171] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.830 [INFO][6171] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.834 [WARNING][6171] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.834 [INFO][6171] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.835 [INFO][6171] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.838808 containerd[1638]: 2026-03-07 01:14:25.837 [INFO][6163] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:25.839178 containerd[1638]: time="2026-03-07T01:14:25.838851679Z" level=info msg="TearDown network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\" successfully" Mar 7 01:14:25.839178 containerd[1638]: time="2026-03-07T01:14:25.838872258Z" level=info msg="StopPodSandbox for \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\" returns successfully" Mar 7 01:14:25.839465 containerd[1638]: time="2026-03-07T01:14:25.839444989Z" level=info msg="RemovePodSandbox for \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\"" Mar 7 01:14:25.839499 containerd[1638]: time="2026-03-07T01:14:25.839467129Z" level=info msg="Forcibly stopping sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\"" Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.861 [WARNING][6185] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"1fd3aff8-915e-4a01-9844-abd28ee38035", ResourceVersion:"1088", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 13, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-6-n-0c8881e772", ContainerID:"03f2cb22e33d77a167eaa82df4a2f9b9d06514a1c01f1e6da3cf1391c425f75e", Pod:"csi-node-driver-xbz8b", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.113.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali394d6dfe0f2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.862 [INFO][6185] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.862 [INFO][6185] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" iface="eth0" netns="" Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.862 [INFO][6185] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.862 [INFO][6185] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.879 [INFO][6193] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.879 [INFO][6193] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.879 [INFO][6193] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.884 [WARNING][6193] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.884 [INFO][6193] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" HandleID="k8s-pod-network.6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Workload="ci--4081--3--6--n--0c8881e772-k8s-csi--node--driver--xbz8b-eth0" Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.885 [INFO][6193] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:14:25.888769 containerd[1638]: 2026-03-07 01:14:25.887 [INFO][6185] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1" Mar 7 01:14:25.889154 containerd[1638]: time="2026-03-07T01:14:25.888787030Z" level=info msg="TearDown network for sandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\" successfully" Mar 7 01:14:25.892605 containerd[1638]: time="2026-03-07T01:14:25.892582138Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:14:25.892658 containerd[1638]: time="2026-03-07T01:14:25.892625576Z" level=info msg="RemovePodSandbox \"6787f86178cf0041540cb34aafa64ba19579a8803cca5151047cedbe46e0bca1\" returns successfully" Mar 7 01:14:27.935035 systemd[1]: run-containerd-runc-k8s.io-4d6f2b013a064a1d00774969e13f724b0eda1d044831eed38b7c44ff377ebbc2-runc.UtkBvh.mount: Deactivated successfully. Mar 7 01:14:44.591248 kubelet[2747]: I0307 01:14:44.590634 2747 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:14:56.052882 systemd[1]: run-containerd-runc-k8s.io-ca2bb687d00bffd48b4dd3b097c611544b1766fe753a5c050a4078a0170f3de3-runc.1cq6p9.mount: Deactivated successfully. Mar 7 01:15:07.715014 update_engine[1622]: I20260307 01:15:07.714880 1622 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Mar 7 01:15:07.715014 update_engine[1622]: I20260307 01:15:07.714973 1622 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Mar 7 01:15:07.715963 update_engine[1622]: I20260307 01:15:07.715417 1622 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Mar 7 01:15:07.717064 update_engine[1622]: I20260307 01:15:07.717005 1622 omaha_request_params.cc:62] Current group set to lts Mar 7 01:15:07.719803 update_engine[1622]: I20260307 01:15:07.718962 1622 update_attempter.cc:499] Already updated boot flags. Skipping. Mar 7 01:15:07.719803 update_engine[1622]: I20260307 01:15:07.718999 1622 update_attempter.cc:643] Scheduling an action processor start. Mar 7 01:15:07.719803 update_engine[1622]: I20260307 01:15:07.719033 1622 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Mar 7 01:15:07.719803 update_engine[1622]: I20260307 01:15:07.719089 1622 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Mar 7 01:15:07.719803 update_engine[1622]: I20260307 01:15:07.719215 1622 omaha_request_action.cc:271] Posting an Omaha request to disabled Mar 7 01:15:07.719803 update_engine[1622]: I20260307 01:15:07.719231 1622 omaha_request_action.cc:272] Request: Mar 7 01:15:07.719803 update_engine[1622]: Mar 7 01:15:07.719803 update_engine[1622]: Mar 7 01:15:07.719803 update_engine[1622]: Mar 7 01:15:07.719803 update_engine[1622]: Mar 7 01:15:07.719803 update_engine[1622]: Mar 7 01:15:07.719803 update_engine[1622]: Mar 7 01:15:07.719803 update_engine[1622]: Mar 7 01:15:07.719803 update_engine[1622]: Mar 7 01:15:07.719803 update_engine[1622]: I20260307 01:15:07.719249 1622 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 7 01:15:07.721572 locksmithd[1663]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Mar 7 01:15:07.725452 update_engine[1622]: I20260307 01:15:07.725395 1622 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 7 01:15:07.726060 update_engine[1622]: I20260307 01:15:07.725983 1622 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 7 01:15:07.726947 update_engine[1622]: E20260307 01:15:07.726883 1622 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 7 01:15:07.727011 update_engine[1622]: I20260307 01:15:07.726991 1622 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Mar 7 01:15:14.149910 systemd[1]: Started sshd@7-95.216.146.133:22-4.153.228.146:57598.service - OpenSSH per-connection server daemon (4.153.228.146:57598). Mar 7 01:15:14.886702 sshd[6362]: Accepted publickey for core from 4.153.228.146 port 57598 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:14.889885 sshd[6362]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:14.897945 systemd-logind[1620]: New session 8 of user core. Mar 7 01:15:14.903006 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 7 01:15:15.501693 sshd[6362]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:15.508947 systemd[1]: sshd@7-95.216.146.133:22-4.153.228.146:57598.service: Deactivated successfully. Mar 7 01:15:15.517536 systemd[1]: session-8.scope: Deactivated successfully. Mar 7 01:15:15.519498 systemd-logind[1620]: Session 8 logged out. Waiting for processes to exit. Mar 7 01:15:15.521428 systemd-logind[1620]: Removed session 8. Mar 7 01:15:17.636830 update_engine[1622]: I20260307 01:15:17.636718 1622 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 7 01:15:17.637605 update_engine[1622]: I20260307 01:15:17.637121 1622 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 7 01:15:17.637605 update_engine[1622]: I20260307 01:15:17.637480 1622 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 7 01:15:17.638262 update_engine[1622]: E20260307 01:15:17.638208 1622 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 7 01:15:17.638329 update_engine[1622]: I20260307 01:15:17.638310 1622 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Mar 7 01:15:20.628053 systemd[1]: Started sshd@8-95.216.146.133:22-4.153.228.146:38754.service - OpenSSH per-connection server daemon (4.153.228.146:38754). Mar 7 01:15:21.378725 sshd[6400]: Accepted publickey for core from 4.153.228.146 port 38754 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:21.381711 sshd[6400]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:21.391047 systemd-logind[1620]: New session 9 of user core. Mar 7 01:15:21.396411 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 7 01:15:21.958738 sshd[6400]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:21.967163 systemd[1]: sshd@8-95.216.146.133:22-4.153.228.146:38754.service: Deactivated successfully. Mar 7 01:15:21.973802 systemd[1]: session-9.scope: Deactivated successfully. Mar 7 01:15:21.976291 systemd-logind[1620]: Session 9 logged out. Waiting for processes to exit. Mar 7 01:15:21.978344 systemd-logind[1620]: Removed session 9. Mar 7 01:15:27.089914 systemd[1]: Started sshd@9-95.216.146.133:22-4.153.228.146:38766.service - OpenSSH per-connection server daemon (4.153.228.146:38766). Mar 7 01:15:27.641884 update_engine[1622]: I20260307 01:15:27.641782 1622 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 7 01:15:27.642681 update_engine[1622]: I20260307 01:15:27.642177 1622 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 7 01:15:27.642681 update_engine[1622]: I20260307 01:15:27.642576 1622 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 7 01:15:27.643279 update_engine[1622]: E20260307 01:15:27.643211 1622 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 7 01:15:27.643330 update_engine[1622]: I20260307 01:15:27.643311 1622 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Mar 7 01:15:27.837198 sshd[6450]: Accepted publickey for core from 4.153.228.146 port 38766 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:27.840192 sshd[6450]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:27.847500 systemd-logind[1620]: New session 10 of user core. Mar 7 01:15:27.854068 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 7 01:15:28.446453 sshd[6450]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:28.452431 systemd[1]: sshd@9-95.216.146.133:22-4.153.228.146:38766.service: Deactivated successfully. Mar 7 01:15:28.462401 systemd[1]: session-10.scope: Deactivated successfully. Mar 7 01:15:28.462997 systemd-logind[1620]: Session 10 logged out. Waiting for processes to exit. Mar 7 01:15:28.465770 systemd-logind[1620]: Removed session 10. Mar 7 01:15:33.572005 systemd[1]: Started sshd@10-95.216.146.133:22-4.153.228.146:49458.service - OpenSSH per-connection server daemon (4.153.228.146:49458). Mar 7 01:15:34.331793 sshd[6516]: Accepted publickey for core from 4.153.228.146 port 49458 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:34.333565 sshd[6516]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:34.340879 systemd-logind[1620]: New session 11 of user core. Mar 7 01:15:34.347803 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 7 01:15:34.952774 sshd[6516]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:34.965085 systemd[1]: sshd@10-95.216.146.133:22-4.153.228.146:49458.service: Deactivated successfully. Mar 7 01:15:34.972865 systemd[1]: session-11.scope: Deactivated successfully. Mar 7 01:15:34.974433 systemd-logind[1620]: Session 11 logged out. Waiting for processes to exit. Mar 7 01:15:34.976933 systemd-logind[1620]: Removed session 11. Mar 7 01:15:35.076634 systemd[1]: Started sshd@11-95.216.146.133:22-4.153.228.146:49474.service - OpenSSH per-connection server daemon (4.153.228.146:49474). Mar 7 01:15:35.813847 sshd[6542]: Accepted publickey for core from 4.153.228.146 port 49474 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:35.815411 sshd[6542]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:35.829910 systemd-logind[1620]: New session 12 of user core. Mar 7 01:15:35.838141 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 7 01:15:36.455163 sshd[6542]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:36.458890 systemd[1]: sshd@11-95.216.146.133:22-4.153.228.146:49474.service: Deactivated successfully. Mar 7 01:15:36.465022 systemd[1]: session-12.scope: Deactivated successfully. Mar 7 01:15:36.465345 systemd-logind[1620]: Session 12 logged out. Waiting for processes to exit. Mar 7 01:15:36.467320 systemd-logind[1620]: Removed session 12. Mar 7 01:15:36.583647 systemd[1]: Started sshd@12-95.216.146.133:22-4.153.228.146:49488.service - OpenSSH per-connection server daemon (4.153.228.146:49488). Mar 7 01:15:37.349396 sshd[6563]: Accepted publickey for core from 4.153.228.146 port 49488 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:37.351616 sshd[6563]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:37.360141 systemd-logind[1620]: New session 13 of user core. Mar 7 01:15:37.366046 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 7 01:15:37.636821 update_engine[1622]: I20260307 01:15:37.636618 1622 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 7 01:15:37.637500 update_engine[1622]: I20260307 01:15:37.637022 1622 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 7 01:15:37.637500 update_engine[1622]: I20260307 01:15:37.637359 1622 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 7 01:15:37.638294 update_engine[1622]: E20260307 01:15:37.638235 1622 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 7 01:15:37.638384 update_engine[1622]: I20260307 01:15:37.638324 1622 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Mar 7 01:15:37.638384 update_engine[1622]: I20260307 01:15:37.638342 1622 omaha_request_action.cc:617] Omaha request response: Mar 7 01:15:37.638517 update_engine[1622]: E20260307 01:15:37.638469 1622 omaha_request_action.cc:636] Omaha request network transfer failed. Mar 7 01:15:37.638593 update_engine[1622]: I20260307 01:15:37.638517 1622 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Mar 7 01:15:37.638593 update_engine[1622]: I20260307 01:15:37.638533 1622 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 7 01:15:37.638678 update_engine[1622]: I20260307 01:15:37.638598 1622 update_attempter.cc:306] Processing Done. Mar 7 01:15:37.638678 update_engine[1622]: E20260307 01:15:37.638624 1622 update_attempter.cc:619] Update failed. Mar 7 01:15:37.638678 update_engine[1622]: I20260307 01:15:37.638639 1622 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Mar 7 01:15:37.638678 update_engine[1622]: I20260307 01:15:37.638654 1622 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Mar 7 01:15:37.638678 update_engine[1622]: I20260307 01:15:37.638669 1622 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Mar 7 01:15:37.638868 update_engine[1622]: I20260307 01:15:37.638773 1622 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Mar 7 01:15:37.638868 update_engine[1622]: I20260307 01:15:37.638808 1622 omaha_request_action.cc:271] Posting an Omaha request to disabled Mar 7 01:15:37.638868 update_engine[1622]: I20260307 01:15:37.638823 1622 omaha_request_action.cc:272] Request: Mar 7 01:15:37.638868 update_engine[1622]: Mar 7 01:15:37.638868 update_engine[1622]: Mar 7 01:15:37.638868 update_engine[1622]: Mar 7 01:15:37.638868 update_engine[1622]: Mar 7 01:15:37.638868 update_engine[1622]: Mar 7 01:15:37.638868 update_engine[1622]: Mar 7 01:15:37.638868 update_engine[1622]: I20260307 01:15:37.638839 1622 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 7 01:15:37.639197 update_engine[1622]: I20260307 01:15:37.639114 1622 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 7 01:15:37.639437 update_engine[1622]: I20260307 01:15:37.639389 1622 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 7 01:15:37.639718 locksmithd[1663]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Mar 7 01:15:37.640594 update_engine[1622]: E20260307 01:15:37.640122 1622 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 7 01:15:37.640594 update_engine[1622]: I20260307 01:15:37.640197 1622 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Mar 7 01:15:37.640594 update_engine[1622]: I20260307 01:15:37.640237 1622 omaha_request_action.cc:617] Omaha request response: Mar 7 01:15:37.640594 update_engine[1622]: I20260307 01:15:37.640254 1622 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 7 01:15:37.640594 update_engine[1622]: I20260307 01:15:37.640268 1622 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 7 01:15:37.640594 update_engine[1622]: I20260307 01:15:37.640283 1622 update_attempter.cc:306] Processing Done. Mar 7 01:15:37.640594 update_engine[1622]: I20260307 01:15:37.640298 1622 update_attempter.cc:310] Error event sent. Mar 7 01:15:37.640594 update_engine[1622]: I20260307 01:15:37.640318 1622 update_check_scheduler.cc:74] Next update check in 42m30s Mar 7 01:15:37.640901 locksmithd[1663]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Mar 7 01:15:37.979137 sshd[6563]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:37.987081 systemd[1]: sshd@12-95.216.146.133:22-4.153.228.146:49488.service: Deactivated successfully. Mar 7 01:15:37.996338 systemd[1]: session-13.scope: Deactivated successfully. Mar 7 01:15:37.998235 systemd-logind[1620]: Session 13 logged out. Waiting for processes to exit. Mar 7 01:15:37.999917 systemd-logind[1620]: Removed session 13. Mar 7 01:15:43.108097 systemd[1]: Started sshd@13-95.216.146.133:22-4.153.228.146:39716.service - OpenSSH per-connection server daemon (4.153.228.146:39716). Mar 7 01:15:43.855220 sshd[6589]: Accepted publickey for core from 4.153.228.146 port 39716 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:43.857506 sshd[6589]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:43.863061 systemd-logind[1620]: New session 14 of user core. Mar 7 01:15:43.869084 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 7 01:15:44.471418 sshd[6589]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:44.476402 systemd[1]: sshd@13-95.216.146.133:22-4.153.228.146:39716.service: Deactivated successfully. Mar 7 01:15:44.484604 systemd-logind[1620]: Session 14 logged out. Waiting for processes to exit. Mar 7 01:15:44.485877 systemd[1]: session-14.scope: Deactivated successfully. Mar 7 01:15:44.488530 systemd-logind[1620]: Removed session 14. Mar 7 01:15:44.597884 systemd[1]: Started sshd@14-95.216.146.133:22-4.153.228.146:39722.service - OpenSSH per-connection server daemon (4.153.228.146:39722). Mar 7 01:15:45.356004 sshd[6621]: Accepted publickey for core from 4.153.228.146 port 39722 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:45.358691 sshd[6621]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:45.364491 systemd-logind[1620]: New session 15 of user core. Mar 7 01:15:45.370076 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 7 01:15:46.131924 sshd[6621]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:46.138426 systemd[1]: sshd@14-95.216.146.133:22-4.153.228.146:39722.service: Deactivated successfully. Mar 7 01:15:46.143291 systemd[1]: session-15.scope: Deactivated successfully. Mar 7 01:15:46.144667 systemd-logind[1620]: Session 15 logged out. Waiting for processes to exit. Mar 7 01:15:46.145966 systemd-logind[1620]: Removed session 15. Mar 7 01:15:46.259107 systemd[1]: Started sshd@15-95.216.146.133:22-4.153.228.146:39724.service - OpenSSH per-connection server daemon (4.153.228.146:39724). Mar 7 01:15:47.010369 sshd[6633]: Accepted publickey for core from 4.153.228.146 port 39724 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:47.013711 sshd[6633]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:47.026746 systemd-logind[1620]: New session 16 of user core. Mar 7 01:15:47.032045 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 7 01:15:47.928963 sshd[6633]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:47.937828 systemd[1]: sshd@15-95.216.146.133:22-4.153.228.146:39724.service: Deactivated successfully. Mar 7 01:15:47.938128 systemd-logind[1620]: Session 16 logged out. Waiting for processes to exit. Mar 7 01:15:47.944869 systemd[1]: session-16.scope: Deactivated successfully. Mar 7 01:15:47.946751 systemd-logind[1620]: Removed session 16. Mar 7 01:15:48.054914 systemd[1]: Started sshd@16-95.216.146.133:22-4.153.228.146:39736.service - OpenSSH per-connection server daemon (4.153.228.146:39736). Mar 7 01:15:48.803749 sshd[6660]: Accepted publickey for core from 4.153.228.146 port 39736 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:48.806479 sshd[6660]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:48.813819 systemd-logind[1620]: New session 17 of user core. Mar 7 01:15:48.816238 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 7 01:15:49.486633 sshd[6660]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:49.491350 systemd[1]: sshd@16-95.216.146.133:22-4.153.228.146:39736.service: Deactivated successfully. Mar 7 01:15:49.501881 systemd-logind[1620]: Session 17 logged out. Waiting for processes to exit. Mar 7 01:15:49.502169 systemd[1]: session-17.scope: Deactivated successfully. Mar 7 01:15:49.505493 systemd-logind[1620]: Removed session 17. Mar 7 01:15:49.616152 systemd[1]: Started sshd@17-95.216.146.133:22-4.153.228.146:51566.service - OpenSSH per-connection server daemon (4.153.228.146:51566). Mar 7 01:15:50.365535 sshd[6692]: Accepted publickey for core from 4.153.228.146 port 51566 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:50.367013 sshd[6692]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:50.373057 systemd-logind[1620]: New session 18 of user core. Mar 7 01:15:50.381155 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 7 01:15:50.976758 sshd[6692]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:50.982724 systemd[1]: sshd@17-95.216.146.133:22-4.153.228.146:51566.service: Deactivated successfully. Mar 7 01:15:50.991693 systemd[1]: session-18.scope: Deactivated successfully. Mar 7 01:15:50.994093 systemd-logind[1620]: Session 18 logged out. Waiting for processes to exit. Mar 7 01:15:50.996450 systemd-logind[1620]: Removed session 18. Mar 7 01:15:56.031813 systemd[1]: run-containerd-runc-k8s.io-ca2bb687d00bffd48b4dd3b097c611544b1766fe753a5c050a4078a0170f3de3-runc.b0r3FD.mount: Deactivated successfully. Mar 7 01:15:56.104186 systemd[1]: Started sshd@18-95.216.146.133:22-4.153.228.146:51582.service - OpenSSH per-connection server daemon (4.153.228.146:51582). Mar 7 01:15:56.858106 sshd[6727]: Accepted publickey for core from 4.153.228.146 port 51582 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:15:56.861412 sshd[6727]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:15:56.868979 systemd-logind[1620]: New session 19 of user core. Mar 7 01:15:56.877076 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 7 01:15:57.476883 sshd[6727]: pam_unix(sshd:session): session closed for user core Mar 7 01:15:57.483757 systemd-logind[1620]: Session 19 logged out. Waiting for processes to exit. Mar 7 01:15:57.485908 systemd[1]: sshd@18-95.216.146.133:22-4.153.228.146:51582.service: Deactivated successfully. Mar 7 01:15:57.493641 systemd[1]: session-19.scope: Deactivated successfully. Mar 7 01:15:57.496817 systemd-logind[1620]: Removed session 19. Mar 7 01:16:02.602574 systemd[1]: Started sshd@19-95.216.146.133:22-4.153.228.146:59504.service - OpenSSH per-connection server daemon (4.153.228.146:59504). Mar 7 01:16:03.357243 sshd[6764]: Accepted publickey for core from 4.153.228.146 port 59504 ssh2: RSA SHA256:cfLbcynJBGQiJlcpT05nBKNU4f9DyADpOV1ay9ga6kI Mar 7 01:16:03.359284 sshd[6764]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:16:03.369659 systemd-logind[1620]: New session 20 of user core. Mar 7 01:16:03.375928 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 7 01:16:03.966929 sshd[6764]: pam_unix(sshd:session): session closed for user core Mar 7 01:16:03.976808 systemd[1]: sshd@19-95.216.146.133:22-4.153.228.146:59504.service: Deactivated successfully. Mar 7 01:16:03.983372 systemd-logind[1620]: Session 20 logged out. Waiting for processes to exit. Mar 7 01:16:03.984397 systemd[1]: session-20.scope: Deactivated successfully. Mar 7 01:16:03.987050 systemd-logind[1620]: Removed session 20. Mar 7 01:16:19.058344 systemd[1]: run-containerd-runc-k8s.io-7c7a8e38c05d4d3d85edb57ad8508f95b57b2e29d482329fc3b28162af264bc0-runc.7uRCCt.mount: Deactivated successfully. Mar 7 01:16:20.939882 kubelet[2747]: E0307 01:16:20.939812 2747 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:36888->10.0.0.2:2379: read: connection timed out" Mar 7 01:16:20.980123 containerd[1638]: time="2026-03-07T01:16:20.979893973Z" level=info msg="shim disconnected" id=cab76e9043845a9f8d584e7d51879f2931b4bee5fe16786c88660c637695e7b4 namespace=k8s.io Mar 7 01:16:20.980744 containerd[1638]: time="2026-03-07T01:16:20.980280792Z" level=warning msg="cleaning up after shim disconnected" id=cab76e9043845a9f8d584e7d51879f2931b4bee5fe16786c88660c637695e7b4 namespace=k8s.io Mar 7 01:16:20.980744 containerd[1638]: time="2026-03-07T01:16:20.980300151Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:16:20.983497 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cab76e9043845a9f8d584e7d51879f2931b4bee5fe16786c88660c637695e7b4-rootfs.mount: Deactivated successfully. Mar 7 01:16:21.247535 containerd[1638]: time="2026-03-07T01:16:21.247081186Z" level=info msg="shim disconnected" id=457a7ea62f0083e306d1f525084462df8f017aa3af4a00716124226e79599df7 namespace=k8s.io Mar 7 01:16:21.247535 containerd[1638]: time="2026-03-07T01:16:21.247157656Z" level=warning msg="cleaning up after shim disconnected" id=457a7ea62f0083e306d1f525084462df8f017aa3af4a00716124226e79599df7 namespace=k8s.io Mar 7 01:16:21.247535 containerd[1638]: time="2026-03-07T01:16:21.247174765Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:16:21.248501 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-457a7ea62f0083e306d1f525084462df8f017aa3af4a00716124226e79599df7-rootfs.mount: Deactivated successfully. Mar 7 01:16:21.596785 kubelet[2747]: I0307 01:16:21.596511 2747 scope.go:117] "RemoveContainer" containerID="cab76e9043845a9f8d584e7d51879f2931b4bee5fe16786c88660c637695e7b4" Mar 7 01:16:21.602075 containerd[1638]: time="2026-03-07T01:16:21.601535815Z" level=info msg="CreateContainer within sandbox \"00c10846e3d6b3c18a74763e0e103d810d570a60b5bd8e0d2ca5c2a955c845a7\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Mar 7 01:16:21.602995 kubelet[2747]: I0307 01:16:21.602787 2747 scope.go:117] "RemoveContainer" containerID="457a7ea62f0083e306d1f525084462df8f017aa3af4a00716124226e79599df7" Mar 7 01:16:21.607176 containerd[1638]: time="2026-03-07T01:16:21.606919003Z" level=info msg="CreateContainer within sandbox \"95bb513489901d098904ca1c6fbe104c8a95276980dad00f918af6d1d85dc0b8\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Mar 7 01:16:21.637673 containerd[1638]: time="2026-03-07T01:16:21.636791331Z" level=info msg="CreateContainer within sandbox \"95bb513489901d098904ca1c6fbe104c8a95276980dad00f918af6d1d85dc0b8\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"c1e2b6475c6f3473a8e3358a290e40c9c1c0aff7cc732af152747ae5d9823d4e\"" Mar 7 01:16:21.638877 containerd[1638]: time="2026-03-07T01:16:21.638759973Z" level=info msg="StartContainer for \"c1e2b6475c6f3473a8e3358a290e40c9c1c0aff7cc732af152747ae5d9823d4e\"" Mar 7 01:16:21.639708 containerd[1638]: time="2026-03-07T01:16:21.639679050Z" level=info msg="CreateContainer within sandbox \"00c10846e3d6b3c18a74763e0e103d810d570a60b5bd8e0d2ca5c2a955c845a7\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"d2724bd2ed77d283729d4f5ee35db5a2915848535e136a343ebceef4891cf94f\"" Mar 7 01:16:21.643771 containerd[1638]: time="2026-03-07T01:16:21.642929427Z" level=info msg="StartContainer for \"d2724bd2ed77d283729d4f5ee35db5a2915848535e136a343ebceef4891cf94f\"" Mar 7 01:16:21.645953 containerd[1638]: time="2026-03-07T01:16:21.645884854Z" level=info msg="shim disconnected" id=3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe namespace=k8s.io Mar 7 01:16:21.646077 containerd[1638]: time="2026-03-07T01:16:21.646056284Z" level=warning msg="cleaning up after shim disconnected" id=3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe namespace=k8s.io Mar 7 01:16:21.646265 containerd[1638]: time="2026-03-07T01:16:21.646214753Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:16:21.648831 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe-rootfs.mount: Deactivated successfully. Mar 7 01:16:21.679862 containerd[1638]: time="2026-03-07T01:16:21.679629497Z" level=warning msg="cleanup warnings time=\"2026-03-07T01:16:21Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Mar 7 01:16:21.732208 containerd[1638]: time="2026-03-07T01:16:21.732162284Z" level=info msg="StartContainer for \"d2724bd2ed77d283729d4f5ee35db5a2915848535e136a343ebceef4891cf94f\" returns successfully" Mar 7 01:16:21.746890 containerd[1638]: time="2026-03-07T01:16:21.746842824Z" level=info msg="StartContainer for \"c1e2b6475c6f3473a8e3358a290e40c9c1c0aff7cc732af152747ae5d9823d4e\" returns successfully" Mar 7 01:16:22.609431 kubelet[2747]: I0307 01:16:22.609365 2747 scope.go:117] "RemoveContainer" containerID="3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe" Mar 7 01:16:22.617017 containerd[1638]: time="2026-03-07T01:16:22.616682983Z" level=info msg="CreateContainer within sandbox \"5bf1d14efe9b97f40fb6b10ebe472dace799d77f722cb43428ea999cb644f370\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Mar 7 01:16:22.642619 containerd[1638]: time="2026-03-07T01:16:22.637218349Z" level=info msg="CreateContainer within sandbox \"5bf1d14efe9b97f40fb6b10ebe472dace799d77f722cb43428ea999cb644f370\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"16fad347de0b57cebff2e64cb5ec95819e6054d4f61b803b3fa27d91162156fb\"" Mar 7 01:16:22.642619 containerd[1638]: time="2026-03-07T01:16:22.638760083Z" level=info msg="StartContainer for \"16fad347de0b57cebff2e64cb5ec95819e6054d4f61b803b3fa27d91162156fb\"" Mar 7 01:16:22.724415 containerd[1638]: time="2026-03-07T01:16:22.724372797Z" level=info msg="StartContainer for \"16fad347de0b57cebff2e64cb5ec95819e6054d4f61b803b3fa27d91162156fb\" returns successfully" Mar 7 01:16:22.982688 systemd[1]: run-containerd-runc-k8s.io-16fad347de0b57cebff2e64cb5ec95819e6054d4f61b803b3fa27d91162156fb-runc.ywLUCD.mount: Deactivated successfully. Mar 7 01:16:24.324108 kubelet[2747]: I0307 01:16:24.324032 2747 status_manager.go:895] "Failed to get status for pod" podUID="4a90aea135cdd8a005d68e1fd7967f4d" pod="kube-system/kube-apiserver-ci-4081-3-6-n-0c8881e772" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:36794->10.0.0.2:2379: read: connection timed out" Mar 7 01:16:25.027720 kubelet[2747]: E0307 01:16:25.027509 2747 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:36686->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4081-3-6-n-0c8881e772.189a6a33eb26dbec kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4081-3-6-n-0c8881e772,UID:4a90aea135cdd8a005d68e1fd7967f4d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4081-3-6-n-0c8881e772,},FirstTimestamp:2026-03-07 01:16:14.544780268 +0000 UTC m=+169.611692025,LastTimestamp:2026-03-07 01:16:14.544780268 +0000 UTC m=+169.611692025,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-6-n-0c8881e772,}" Mar 7 01:16:25.226502 systemd[1]: run-containerd-runc-k8s.io-7c7a8e38c05d4d3d85edb57ad8508f95b57b2e29d482329fc3b28162af264bc0-runc.03qbyq.mount: Deactivated successfully. Mar 7 01:16:30.940751 kubelet[2747]: E0307 01:16:30.940569 2747 controller.go:195] "Failed to update lease" err="Put \"https://95.216.146.133:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-0c8881e772?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 7 01:16:33.960051 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-16fad347de0b57cebff2e64cb5ec95819e6054d4f61b803b3fa27d91162156fb-rootfs.mount: Deactivated successfully. Mar 7 01:16:33.969888 containerd[1638]: time="2026-03-07T01:16:33.969765900Z" level=info msg="shim disconnected" id=16fad347de0b57cebff2e64cb5ec95819e6054d4f61b803b3fa27d91162156fb namespace=k8s.io Mar 7 01:16:33.969888 containerd[1638]: time="2026-03-07T01:16:33.969834360Z" level=warning msg="cleaning up after shim disconnected" id=16fad347de0b57cebff2e64cb5ec95819e6054d4f61b803b3fa27d91162156fb namespace=k8s.io Mar 7 01:16:33.969888 containerd[1638]: time="2026-03-07T01:16:33.969850100Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:16:34.656279 kubelet[2747]: I0307 01:16:34.656218 2747 scope.go:117] "RemoveContainer" containerID="3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe" Mar 7 01:16:34.657128 kubelet[2747]: I0307 01:16:34.656798 2747 scope.go:117] "RemoveContainer" containerID="16fad347de0b57cebff2e64cb5ec95819e6054d4f61b803b3fa27d91162156fb" Mar 7 01:16:34.657128 kubelet[2747]: E0307 01:16:34.657015 2747 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tigera-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=tigera-operator pod=tigera-operator-6bf85f8dd-ljzps_tigera-operator(7e4a93f4-9c0e-44f9-bb3a-dcba317b694e)\"" pod="tigera-operator/tigera-operator-6bf85f8dd-ljzps" podUID="7e4a93f4-9c0e-44f9-bb3a-dcba317b694e" Mar 7 01:16:34.659155 containerd[1638]: time="2026-03-07T01:16:34.659065031Z" level=info msg="RemoveContainer for \"3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe\"" Mar 7 01:16:34.666299 containerd[1638]: time="2026-03-07T01:16:34.666223867Z" level=info msg="RemoveContainer for \"3313ca3e5ae4bec0ed8e3249c86b24a897a35a0a44e4ce9a146e3ec13b628fbe\" returns successfully" Mar 7 01:16:40.942394 kubelet[2747]: E0307 01:16:40.942324 2747 controller.go:195] "Failed to update lease" err="Put \"https://95.216.146.133:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-0c8881e772?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"