Mar 6 02:33:43.112252 kernel: Linux version 6.12.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Thu Mar 5 23:16:40 -00 2026 Mar 6 02:33:43.112659 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=5bef16c10382b6f77f9493af2297475832ff2f09f1ada4155425ad9b32dd6e53 Mar 6 02:33:43.112673 kernel: BIOS-provided physical RAM map: Mar 6 02:33:43.112687 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 6 02:33:43.112696 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Mar 6 02:33:43.112705 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Mar 6 02:33:43.112715 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Mar 6 02:33:43.112724 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Mar 6 02:33:43.112733 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Mar 6 02:33:43.112943 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Mar 6 02:33:43.112955 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Mar 6 02:33:43.112965 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Mar 6 02:33:43.112978 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Mar 6 02:33:43.112987 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Mar 6 02:33:43.112996 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Mar 6 02:33:43.113005 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Mar 6 02:33:43.113013 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Mar 6 02:33:43.113025 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Mar 6 02:33:43.113033 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Mar 6 02:33:43.113042 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Mar 6 02:33:43.113052 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Mar 6 02:33:43.113061 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Mar 6 02:33:43.113071 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Mar 6 02:33:43.113081 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 6 02:33:43.113090 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Mar 6 02:33:43.113100 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 6 02:33:43.113109 kernel: NX (Execute Disable) protection: active Mar 6 02:33:43.113119 kernel: APIC: Static calls initialized Mar 6 02:33:43.113132 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Mar 6 02:33:43.113142 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Mar 6 02:33:43.113152 kernel: extended physical RAM map: Mar 6 02:33:43.113162 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 6 02:33:43.113171 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Mar 6 02:33:43.113181 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Mar 6 02:33:43.113191 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Mar 6 02:33:43.113201 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Mar 6 02:33:43.113210 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Mar 6 02:33:43.113221 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Mar 6 02:33:43.113231 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Mar 6 02:33:43.113244 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Mar 6 02:33:43.113258 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Mar 6 02:33:43.113269 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Mar 6 02:33:43.113279 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Mar 6 02:33:43.113406 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Mar 6 02:33:43.113421 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Mar 6 02:33:43.113432 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Mar 6 02:33:43.113442 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Mar 6 02:33:43.113452 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Mar 6 02:33:43.113463 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Mar 6 02:33:43.113473 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Mar 6 02:33:43.113483 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Mar 6 02:33:43.113493 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Mar 6 02:33:43.113503 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Mar 6 02:33:43.113514 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Mar 6 02:33:43.113524 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Mar 6 02:33:43.113538 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 6 02:33:43.113549 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Mar 6 02:33:43.113559 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 6 02:33:43.113569 kernel: efi: EFI v2.7 by EDK II Mar 6 02:33:43.113580 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Mar 6 02:33:43.113591 kernel: random: crng init done Mar 6 02:33:43.113601 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Mar 6 02:33:43.113611 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Mar 6 02:33:43.113622 kernel: secureboot: Secure boot disabled Mar 6 02:33:43.113632 kernel: SMBIOS 2.8 present. Mar 6 02:33:43.113642 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Mar 6 02:33:43.113656 kernel: DMI: Memory slots populated: 1/1 Mar 6 02:33:43.113666 kernel: Hypervisor detected: KVM Mar 6 02:33:43.113677 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Mar 6 02:33:43.113687 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 6 02:33:43.113697 kernel: kvm-clock: using sched offset of 18476242941 cycles Mar 6 02:33:43.113708 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 6 02:33:43.113719 kernel: tsc: Detected 2445.424 MHz processor Mar 6 02:33:43.113729 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 6 02:33:43.113949 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 6 02:33:43.113966 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Mar 6 02:33:43.113977 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Mar 6 02:33:43.113991 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 6 02:33:43.114001 kernel: Using GB pages for direct mapping Mar 6 02:33:43.114010 kernel: ACPI: Early table checksum verification disabled Mar 6 02:33:43.114019 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Mar 6 02:33:43.114028 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Mar 6 02:33:43.114038 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 02:33:43.114047 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 02:33:43.114057 kernel: ACPI: FACS 0x000000009CBDD000 000040 Mar 6 02:33:43.114067 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 02:33:43.114081 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 02:33:43.114092 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 02:33:43.114102 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 02:33:43.114113 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Mar 6 02:33:43.114123 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Mar 6 02:33:43.114133 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Mar 6 02:33:43.114144 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Mar 6 02:33:43.114154 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Mar 6 02:33:43.114168 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Mar 6 02:33:43.114179 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Mar 6 02:33:43.114190 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Mar 6 02:33:43.114200 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Mar 6 02:33:43.114211 kernel: No NUMA configuration found Mar 6 02:33:43.114222 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Mar 6 02:33:43.114232 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Mar 6 02:33:43.114243 kernel: Zone ranges: Mar 6 02:33:43.114254 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 6 02:33:43.114267 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Mar 6 02:33:43.114277 kernel: Normal empty Mar 6 02:33:43.114406 kernel: Device empty Mar 6 02:33:43.114418 kernel: Movable zone start for each node Mar 6 02:33:43.114429 kernel: Early memory node ranges Mar 6 02:33:43.114440 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Mar 6 02:33:43.114451 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Mar 6 02:33:43.114462 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Mar 6 02:33:43.114472 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Mar 6 02:33:43.114484 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Mar 6 02:33:43.114500 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Mar 6 02:33:43.114511 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Mar 6 02:33:43.114522 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Mar 6 02:33:43.114533 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Mar 6 02:33:43.114544 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 6 02:33:43.114567 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Mar 6 02:33:43.114582 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Mar 6 02:33:43.114593 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 6 02:33:43.114605 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Mar 6 02:33:43.114616 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Mar 6 02:33:43.114628 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Mar 6 02:33:43.114639 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Mar 6 02:33:43.114654 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Mar 6 02:33:43.114665 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 6 02:33:43.114676 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 6 02:33:43.114687 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 6 02:33:43.114699 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 6 02:33:43.114714 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 6 02:33:43.114725 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 6 02:33:43.114736 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 6 02:33:43.114971 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 6 02:33:43.114984 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 6 02:33:43.114993 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 6 02:33:43.115003 kernel: TSC deadline timer available Mar 6 02:33:43.115012 kernel: CPU topo: Max. logical packages: 1 Mar 6 02:33:43.115022 kernel: CPU topo: Max. logical dies: 1 Mar 6 02:33:43.115036 kernel: CPU topo: Max. dies per package: 1 Mar 6 02:33:43.115045 kernel: CPU topo: Max. threads per core: 1 Mar 6 02:33:43.115055 kernel: CPU topo: Num. cores per package: 4 Mar 6 02:33:43.115065 kernel: CPU topo: Num. threads per package: 4 Mar 6 02:33:43.115076 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Mar 6 02:33:43.115087 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 6 02:33:43.115098 kernel: kvm-guest: KVM setup pv remote TLB flush Mar 6 02:33:43.115109 kernel: kvm-guest: setup PV sched yield Mar 6 02:33:43.115120 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Mar 6 02:33:43.115136 kernel: Booting paravirtualized kernel on KVM Mar 6 02:33:43.115147 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 6 02:33:43.115159 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Mar 6 02:33:43.115170 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Mar 6 02:33:43.115181 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Mar 6 02:33:43.115192 kernel: pcpu-alloc: [0] 0 1 2 3 Mar 6 02:33:43.115204 kernel: kvm-guest: PV spinlocks enabled Mar 6 02:33:43.115216 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Mar 6 02:33:43.115229 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=5bef16c10382b6f77f9493af2297475832ff2f09f1ada4155425ad9b32dd6e53 Mar 6 02:33:43.115246 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 6 02:33:43.115257 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 6 02:33:43.115269 kernel: Fallback order for Node 0: 0 Mar 6 02:33:43.115280 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Mar 6 02:33:43.115410 kernel: Policy zone: DMA32 Mar 6 02:33:43.115422 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 6 02:33:43.115433 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Mar 6 02:33:43.115443 kernel: ftrace: allocating 40099 entries in 157 pages Mar 6 02:33:43.115459 kernel: ftrace: allocated 157 pages with 5 groups Mar 6 02:33:43.115470 kernel: Dynamic Preempt: voluntary Mar 6 02:33:43.115481 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 6 02:33:43.115493 kernel: rcu: RCU event tracing is enabled. Mar 6 02:33:43.115505 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Mar 6 02:33:43.115516 kernel: Trampoline variant of Tasks RCU enabled. Mar 6 02:33:43.115528 kernel: Rude variant of Tasks RCU enabled. Mar 6 02:33:43.115539 kernel: Tracing variant of Tasks RCU enabled. Mar 6 02:33:43.115551 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 6 02:33:43.115562 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Mar 6 02:33:43.115578 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 6 02:33:43.115589 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 6 02:33:43.115601 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 6 02:33:43.115612 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Mar 6 02:33:43.115623 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 6 02:33:43.115635 kernel: Console: colour dummy device 80x25 Mar 6 02:33:43.115646 kernel: printk: legacy console [ttyS0] enabled Mar 6 02:33:43.115657 kernel: ACPI: Core revision 20240827 Mar 6 02:33:43.115669 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 6 02:33:43.115684 kernel: APIC: Switch to symmetric I/O mode setup Mar 6 02:33:43.115695 kernel: x2apic enabled Mar 6 02:33:43.115706 kernel: APIC: Switched APIC routing to: physical x2apic Mar 6 02:33:43.115718 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Mar 6 02:33:43.115729 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Mar 6 02:33:43.116049 kernel: kvm-guest: setup PV IPIs Mar 6 02:33:43.116064 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 6 02:33:43.116074 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fd5e8294, max_idle_ns: 440795237246 ns Mar 6 02:33:43.116084 kernel: Calibrating delay loop (skipped) preset value.. 4890.84 BogoMIPS (lpj=2445424) Mar 6 02:33:43.116098 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 6 02:33:43.116108 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Mar 6 02:33:43.116118 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Mar 6 02:33:43.116127 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 6 02:33:43.116137 kernel: Spectre V2 : Mitigation: Retpolines Mar 6 02:33:43.116146 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Mar 6 02:33:43.116157 kernel: Speculative Store Bypass: Vulnerable Mar 6 02:33:43.116168 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Mar 6 02:33:43.116185 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Mar 6 02:33:43.116197 kernel: active return thunk: srso_alias_return_thunk Mar 6 02:33:43.116210 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Mar 6 02:33:43.116221 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Mar 6 02:33:43.116233 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Mar 6 02:33:43.116245 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 6 02:33:43.116257 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 6 02:33:43.116269 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 6 02:33:43.116280 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 6 02:33:43.116435 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Mar 6 02:33:43.116447 kernel: Freeing SMP alternatives memory: 32K Mar 6 02:33:43.116459 kernel: pid_max: default: 32768 minimum: 301 Mar 6 02:33:43.116470 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Mar 6 02:33:43.116481 kernel: landlock: Up and running. Mar 6 02:33:43.116493 kernel: SELinux: Initializing. Mar 6 02:33:43.116504 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 6 02:33:43.116515 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 6 02:33:43.116526 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Mar 6 02:33:43.116542 kernel: Performance Events: PMU not available due to virtualization, using software events only. Mar 6 02:33:43.116554 kernel: signal: max sigframe size: 1776 Mar 6 02:33:43.116566 kernel: rcu: Hierarchical SRCU implementation. Mar 6 02:33:43.116579 kernel: rcu: Max phase no-delay instances is 400. Mar 6 02:33:43.116591 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Mar 6 02:33:43.116603 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Mar 6 02:33:43.116615 kernel: smp: Bringing up secondary CPUs ... Mar 6 02:33:43.116626 kernel: smpboot: x86: Booting SMP configuration: Mar 6 02:33:43.116638 kernel: .... node #0, CPUs: #1 #2 #3 Mar 6 02:33:43.116653 kernel: smp: Brought up 1 node, 4 CPUs Mar 6 02:33:43.116665 kernel: smpboot: Total of 4 processors activated (19563.39 BogoMIPS) Mar 6 02:33:43.116678 kernel: Memory: 2414472K/2565800K available (14336K kernel code, 2445K rwdata, 26064K rodata, 46196K init, 2564K bss, 145388K reserved, 0K cma-reserved) Mar 6 02:33:43.116689 kernel: devtmpfs: initialized Mar 6 02:33:43.116701 kernel: x86/mm: Memory block size: 128MB Mar 6 02:33:43.116713 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Mar 6 02:33:43.116724 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Mar 6 02:33:43.116736 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Mar 6 02:33:43.117078 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Mar 6 02:33:43.117091 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Mar 6 02:33:43.117101 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Mar 6 02:33:43.117111 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 6 02:33:43.117121 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Mar 6 02:33:43.117130 kernel: pinctrl core: initialized pinctrl subsystem Mar 6 02:33:43.117140 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 6 02:33:43.117150 kernel: audit: initializing netlink subsys (disabled) Mar 6 02:33:43.117159 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 6 02:33:43.117173 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 6 02:33:43.117185 kernel: audit: type=2000 audit(1772764391.554:1): state=initialized audit_enabled=0 res=1 Mar 6 02:33:43.117195 kernel: cpuidle: using governor menu Mar 6 02:33:43.117206 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 6 02:33:43.117217 kernel: dca service started, version 1.12.1 Mar 6 02:33:43.117228 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Mar 6 02:33:43.117240 kernel: PCI: Using configuration type 1 for base access Mar 6 02:33:43.117251 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 6 02:33:43.117262 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 6 02:33:43.117277 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 6 02:33:43.117445 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 6 02:33:43.117458 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 6 02:33:43.117469 kernel: ACPI: Added _OSI(Module Device) Mar 6 02:33:43.117480 kernel: ACPI: Added _OSI(Processor Device) Mar 6 02:33:43.117491 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 6 02:33:43.117502 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 6 02:33:43.117513 kernel: ACPI: Interpreter enabled Mar 6 02:33:43.117525 kernel: ACPI: PM: (supports S0 S3 S5) Mar 6 02:33:43.117540 kernel: ACPI: Using IOAPIC for interrupt routing Mar 6 02:33:43.117551 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 6 02:33:43.117562 kernel: PCI: Using E820 reservations for host bridge windows Mar 6 02:33:43.117573 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 6 02:33:43.117584 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 6 02:33:43.123699 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 6 02:33:43.124589 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Mar 6 02:33:43.125007 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Mar 6 02:33:43.125031 kernel: PCI host bridge to bus 0000:00 Mar 6 02:33:43.125210 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 6 02:33:43.125510 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 6 02:33:43.125691 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 6 02:33:43.126086 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Mar 6 02:33:43.126492 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Mar 6 02:33:43.126676 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Mar 6 02:33:43.127070 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 6 02:33:43.127280 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Mar 6 02:33:43.128403 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Mar 6 02:33:43.128983 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Mar 6 02:33:43.129272 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Mar 6 02:33:43.129572 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Mar 6 02:33:43.129983 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 6 02:33:43.130165 kernel: pci 0000:00:01.0: pci_fixup_video+0x0/0x100 took 18554 usecs Mar 6 02:33:43.130489 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Mar 6 02:33:43.130673 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Mar 6 02:33:43.131083 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Mar 6 02:33:43.131402 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Mar 6 02:33:43.131613 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Mar 6 02:33:43.132037 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Mar 6 02:33:43.132236 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Mar 6 02:33:43.132545 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Mar 6 02:33:43.132986 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Mar 6 02:33:43.133176 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Mar 6 02:33:43.133499 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Mar 6 02:33:43.133706 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Mar 6 02:33:43.134190 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Mar 6 02:33:43.134520 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Mar 6 02:33:43.134705 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 6 02:33:43.135239 kernel: pci 0000:00:1f.0: quirk_ich7_lpc+0x0/0xc0 took 15625 usecs Mar 6 02:33:43.135564 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Mar 6 02:33:43.135976 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Mar 6 02:33:43.136165 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Mar 6 02:33:43.136485 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Mar 6 02:33:43.136670 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Mar 6 02:33:43.136689 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 6 02:33:43.136702 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 6 02:33:43.136715 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 6 02:33:43.136727 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 6 02:33:43.136959 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 6 02:33:43.136984 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 6 02:33:43.136995 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 6 02:33:43.137004 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 6 02:33:43.137014 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 6 02:33:43.137023 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 6 02:33:43.137033 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 6 02:33:43.137043 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 6 02:33:43.137053 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 6 02:33:43.137063 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 6 02:33:43.137079 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 6 02:33:43.137091 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 6 02:33:43.137103 kernel: iommu: Default domain type: Translated Mar 6 02:33:43.137115 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 6 02:33:43.137126 kernel: efivars: Registered efivars operations Mar 6 02:33:43.137138 kernel: PCI: Using ACPI for IRQ routing Mar 6 02:33:43.137149 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 6 02:33:43.137161 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Mar 6 02:33:43.137172 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Mar 6 02:33:43.137188 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Mar 6 02:33:43.137199 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Mar 6 02:33:43.137210 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Mar 6 02:33:43.137223 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Mar 6 02:33:43.137233 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Mar 6 02:33:43.137243 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Mar 6 02:33:43.137539 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 6 02:33:43.137724 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 6 02:33:43.138245 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 6 02:33:43.138264 kernel: vgaarb: loaded Mar 6 02:33:43.138276 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 6 02:33:43.138407 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 6 02:33:43.138420 kernel: clocksource: Switched to clocksource kvm-clock Mar 6 02:33:43.138432 kernel: VFS: Disk quotas dquot_6.6.0 Mar 6 02:33:43.138443 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 6 02:33:43.138454 kernel: pnp: PnP ACPI init Mar 6 02:33:43.138650 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Mar 6 02:33:43.138675 kernel: pnp: PnP ACPI: found 6 devices Mar 6 02:33:43.138687 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 6 02:33:43.138699 kernel: NET: Registered PF_INET protocol family Mar 6 02:33:43.138710 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 6 02:33:43.138722 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 6 02:33:43.138981 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 6 02:33:43.138997 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 6 02:33:43.139008 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 6 02:33:43.139021 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 6 02:33:43.139031 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 6 02:33:43.139041 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 6 02:33:43.139051 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 6 02:33:43.139061 kernel: NET: Registered PF_XDP protocol family Mar 6 02:33:43.139243 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Mar 6 02:33:43.139546 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Mar 6 02:33:43.140122 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 6 02:33:43.140422 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 6 02:33:43.140594 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 6 02:33:43.140988 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Mar 6 02:33:43.141152 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Mar 6 02:33:43.141433 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Mar 6 02:33:43.141453 kernel: PCI: CLS 0 bytes, default 64 Mar 6 02:33:43.141466 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fd5e8294, max_idle_ns: 440795237246 ns Mar 6 02:33:43.141479 kernel: Initialise system trusted keyrings Mar 6 02:33:43.141491 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 6 02:33:43.141508 kernel: Key type asymmetric registered Mar 6 02:33:43.141521 kernel: Asymmetric key parser 'x509' registered Mar 6 02:33:43.141533 kernel: hrtimer: interrupt took 6310073 ns Mar 6 02:33:43.141546 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Mar 6 02:33:43.141559 kernel: io scheduler mq-deadline registered Mar 6 02:33:43.141571 kernel: io scheduler kyber registered Mar 6 02:33:43.141583 kernel: io scheduler bfq registered Mar 6 02:33:43.141599 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 6 02:33:43.141613 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 6 02:33:43.141628 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 6 02:33:43.141641 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Mar 6 02:33:43.141652 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 6 02:33:43.141664 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 6 02:33:43.141676 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 6 02:33:43.141692 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 6 02:33:43.141704 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 6 02:33:43.142118 kernel: rtc_cmos 00:04: RTC can wake from S4 Mar 6 02:33:43.142138 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1 Mar 6 02:33:43.142433 kernel: rtc_cmos 00:04: registered as rtc0 Mar 6 02:33:43.142610 kernel: rtc_cmos 00:04: setting system clock to 2026-03-06T02:33:36 UTC (1772764416) Mar 6 02:33:43.143187 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Mar 6 02:33:43.143210 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Mar 6 02:33:43.143229 kernel: efifb: probing for efifb Mar 6 02:33:43.143242 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Mar 6 02:33:43.143258 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Mar 6 02:33:43.143271 kernel: efifb: scrolling: redraw Mar 6 02:33:43.143400 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 6 02:33:43.143416 kernel: Console: switching to colour frame buffer device 160x50 Mar 6 02:33:43.143429 kernel: fb0: EFI VGA frame buffer device Mar 6 02:33:43.143441 kernel: pstore: Using crash dump compression: deflate Mar 6 02:33:43.143453 kernel: pstore: Registered efi_pstore as persistent store backend Mar 6 02:33:43.143469 kernel: NET: Registered PF_INET6 protocol family Mar 6 02:33:43.143480 kernel: Segment Routing with IPv6 Mar 6 02:33:43.143493 kernel: In-situ OAM (IOAM) with IPv6 Mar 6 02:33:43.143505 kernel: NET: Registered PF_PACKET protocol family Mar 6 02:33:43.143517 kernel: Key type dns_resolver registered Mar 6 02:33:43.143530 kernel: IPI shorthand broadcast: enabled Mar 6 02:33:43.143542 kernel: sched_clock: Marking stable (23562130452, 769912682)->(27225304986, -2893261852) Mar 6 02:33:43.143555 kernel: registered taskstats version 1 Mar 6 02:33:43.143566 kernel: Loading compiled-in X.509 certificates Mar 6 02:33:43.143578 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.74-flatcar: 30893fe9fd219d26109af079e6493e1c8b1c00af' Mar 6 02:33:43.143594 kernel: Demotion targets for Node 0: null Mar 6 02:33:43.143606 kernel: Key type .fscrypt registered Mar 6 02:33:43.143617 kernel: Key type fscrypt-provisioning registered Mar 6 02:33:43.143629 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 6 02:33:43.143641 kernel: ima: Allocated hash algorithm: sha1 Mar 6 02:33:43.143653 kernel: ima: No architecture policies found Mar 6 02:33:43.143666 kernel: clk: Disabling unused clocks Mar 6 02:33:43.143678 kernel: Warning: unable to open an initial console. Mar 6 02:33:43.143693 kernel: Freeing unused kernel image (initmem) memory: 46196K Mar 6 02:33:43.143705 kernel: Write protecting the kernel read-only data: 40960k Mar 6 02:33:43.143717 kernel: Freeing unused kernel image (rodata/data gap) memory: 560K Mar 6 02:33:43.143729 kernel: Run /init as init process Mar 6 02:33:43.144061 kernel: with arguments: Mar 6 02:33:43.144078 kernel: /init Mar 6 02:33:43.144091 kernel: with environment: Mar 6 02:33:43.144103 kernel: HOME=/ Mar 6 02:33:43.144115 kernel: TERM=linux Mar 6 02:33:43.144135 systemd[1]: Successfully made /usr/ read-only. Mar 6 02:33:43.144153 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 6 02:33:43.144167 systemd[1]: Detected virtualization kvm. Mar 6 02:33:43.144179 systemd[1]: Detected architecture x86-64. Mar 6 02:33:43.144192 systemd[1]: Running in initrd. Mar 6 02:33:43.144204 systemd[1]: No hostname configured, using default hostname. Mar 6 02:33:43.144218 systemd[1]: Hostname set to . Mar 6 02:33:43.144234 systemd[1]: Initializing machine ID from VM UUID. Mar 6 02:33:43.144246 systemd[1]: Queued start job for default target initrd.target. Mar 6 02:33:43.144259 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 6 02:33:43.144272 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 6 02:33:43.144400 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 6 02:33:43.144416 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 6 02:33:43.144429 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 6 02:33:43.144443 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 6 02:33:43.144461 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 6 02:33:43.144475 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 6 02:33:43.144487 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 6 02:33:43.144500 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 6 02:33:43.144512 systemd[1]: Reached target paths.target - Path Units. Mar 6 02:33:43.144525 systemd[1]: Reached target slices.target - Slice Units. Mar 6 02:33:43.144537 systemd[1]: Reached target swap.target - Swaps. Mar 6 02:33:43.144553 systemd[1]: Reached target timers.target - Timer Units. Mar 6 02:33:43.144566 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 6 02:33:43.144579 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 6 02:33:43.144591 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 6 02:33:43.144604 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Mar 6 02:33:43.144617 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 6 02:33:43.144630 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 6 02:33:43.144643 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 6 02:33:43.144656 systemd[1]: Reached target sockets.target - Socket Units. Mar 6 02:33:43.144673 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 6 02:33:43.144685 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 6 02:33:43.144698 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 6 02:33:43.144710 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Mar 6 02:33:43.144722 systemd[1]: Starting systemd-fsck-usr.service... Mar 6 02:33:43.144735 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 6 02:33:43.144972 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 6 02:33:43.144986 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 02:33:43.145001 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 6 02:33:43.145013 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 6 02:33:43.145023 systemd[1]: Finished systemd-fsck-usr.service. Mar 6 02:33:43.145179 systemd-journald[204]: Collecting audit messages is disabled. Mar 6 02:33:43.145220 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 6 02:33:43.145235 systemd-journald[204]: Journal started Mar 6 02:33:43.145490 systemd-journald[204]: Runtime Journal (/run/log/journal/86b8fd27b29f4970962f2d2e02c7221e) is 6M, max 48.1M, 42.1M free. Mar 6 02:33:43.075060 systemd-modules-load[205]: Inserted module 'overlay' Mar 6 02:33:43.227427 systemd[1]: Started systemd-journald.service - Journal Service. Mar 6 02:33:43.311235 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 02:33:43.366125 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 6 02:33:43.395416 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 6 02:33:43.594223 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 6 02:33:43.638510 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 6 02:33:43.652725 systemd-tmpfiles[220]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Mar 6 02:33:43.720178 kernel: Bridge firewalling registered Mar 6 02:33:43.666533 systemd-modules-load[205]: Inserted module 'br_netfilter' Mar 6 02:33:43.669210 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 6 02:33:43.687564 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 6 02:33:43.723601 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 6 02:33:43.825994 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 6 02:33:43.860048 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 6 02:33:43.942449 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 6 02:33:44.075238 dracut-cmdline[235]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=5bef16c10382b6f77f9493af2297475832ff2f09f1ada4155425ad9b32dd6e53 Mar 6 02:33:44.091643 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 6 02:33:44.206138 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 6 02:33:44.300124 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 6 02:33:44.542945 systemd-resolved[269]: Positive Trust Anchors: Mar 6 02:33:44.543063 systemd-resolved[269]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 6 02:33:44.543104 systemd-resolved[269]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 6 02:33:44.554081 systemd-resolved[269]: Defaulting to hostname 'linux'. Mar 6 02:33:44.559523 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 6 02:33:44.639934 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 6 02:33:44.823863 kernel: SCSI subsystem initialized Mar 6 02:33:44.843116 kernel: Loading iSCSI transport class v2.0-870. Mar 6 02:33:44.881525 kernel: iscsi: registered transport (tcp) Mar 6 02:33:44.942698 kernel: iscsi: registered transport (qla4xxx) Mar 6 02:33:44.942989 kernel: QLogic iSCSI HBA Driver Mar 6 02:33:45.003174 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 6 02:33:45.063000 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 6 02:33:45.075900 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 6 02:33:45.264640 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 6 02:33:45.283074 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 6 02:33:45.438082 kernel: raid6: avx2x4 gen() 7605 MB/s Mar 6 02:33:45.458476 kernel: raid6: avx2x2 gen() 14916 MB/s Mar 6 02:33:45.480161 kernel: raid6: avx2x1 gen() 16850 MB/s Mar 6 02:33:45.480241 kernel: raid6: using algorithm avx2x1 gen() 16850 MB/s Mar 6 02:33:45.502983 kernel: raid6: .... xor() 18059 MB/s, rmw enabled Mar 6 02:33:45.503061 kernel: raid6: using avx2x2 recovery algorithm Mar 6 02:33:45.539946 kernel: xor: automatically using best checksumming function avx Mar 6 02:33:46.601646 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 6 02:33:46.650983 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 6 02:33:46.674389 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 6 02:33:46.777190 systemd-udevd[454]: Using default interface naming scheme 'v255'. Mar 6 02:33:46.792208 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 6 02:33:46.821064 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 6 02:33:46.952691 dracut-pre-trigger[458]: rd.md=0: removing MD RAID activation Mar 6 02:33:47.059658 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 6 02:33:47.064911 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 6 02:33:47.310215 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 6 02:33:47.339631 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 6 02:33:47.502037 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Mar 6 02:33:47.502423 kernel: cryptd: max_cpu_qlen set to 1000 Mar 6 02:33:47.543036 kernel: libata version 3.00 loaded. Mar 6 02:33:47.562080 kernel: AES CTR mode by8 optimization enabled Mar 6 02:33:47.569010 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 6 02:33:47.569604 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 02:33:47.595864 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Mar 6 02:33:47.609041 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 02:33:47.683180 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 6 02:33:47.683208 kernel: GPT:9289727 != 19775487 Mar 6 02:33:47.683219 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 6 02:33:47.683228 kernel: GPT:9289727 != 19775487 Mar 6 02:33:47.683237 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 6 02:33:47.683247 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 6 02:33:47.683256 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Mar 6 02:33:47.683266 kernel: ahci 0000:00:1f.2: version 3.0 Mar 6 02:33:47.683602 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 6 02:33:47.658184 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 02:33:47.713572 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Mar 6 02:33:47.714547 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Mar 6 02:33:47.714911 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 6 02:33:47.724527 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 6 02:33:47.742223 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 6 02:33:47.764938 kernel: scsi host0: ahci Mar 6 02:33:47.765230 kernel: scsi host1: ahci Mar 6 02:33:47.746523 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 02:33:47.790934 kernel: scsi host2: ahci Mar 6 02:33:47.796986 kernel: scsi host3: ahci Mar 6 02:33:47.801968 kernel: scsi host4: ahci Mar 6 02:33:47.809240 kernel: scsi host5: ahci Mar 6 02:33:47.825043 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 31 lpm-pol 1 Mar 6 02:33:47.825112 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 31 lpm-pol 1 Mar 6 02:33:47.825145 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 31 lpm-pol 1 Mar 6 02:33:47.834679 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 6 02:33:47.870213 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 31 lpm-pol 1 Mar 6 02:33:47.870249 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 31 lpm-pol 1 Mar 6 02:33:47.870265 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 31 lpm-pol 1 Mar 6 02:33:47.865157 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 6 02:33:47.892719 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Mar 6 02:33:47.902994 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 6 02:33:47.935568 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 6 02:33:47.941648 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 6 02:33:47.952050 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 02:33:47.998100 disk-uuid[617]: Primary Header is updated. Mar 6 02:33:47.998100 disk-uuid[617]: Secondary Entries is updated. Mar 6 02:33:47.998100 disk-uuid[617]: Secondary Header is updated. Mar 6 02:33:48.022952 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 6 02:33:48.025636 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 02:33:48.166336 kernel: ata1: SATA link down (SStatus 0 SControl 300) Mar 6 02:33:48.196373 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Mar 6 02:33:48.196445 kernel: ata3.00: LPM support broken, forcing max_power Mar 6 02:33:48.196463 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Mar 6 02:33:48.196478 kernel: ata3.00: applying bridge limits Mar 6 02:33:48.206948 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 6 02:33:48.213881 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 6 02:33:48.213937 kernel: ata3.00: LPM support broken, forcing max_power Mar 6 02:33:48.221970 kernel: ata3.00: configured for UDMA/100 Mar 6 02:33:48.232896 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 6 02:33:48.238196 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 6 02:33:48.242950 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 6 02:33:48.295956 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Mar 6 02:33:48.296350 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 6 02:33:48.316972 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Mar 6 02:33:48.825642 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 6 02:33:48.837234 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 6 02:33:48.854461 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 6 02:33:48.899728 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 6 02:33:48.919944 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 6 02:33:48.976669 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 6 02:33:49.052016 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 6 02:33:49.054840 disk-uuid[620]: The operation has completed successfully. Mar 6 02:33:49.138952 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 6 02:33:49.139241 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 6 02:33:49.198633 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 6 02:33:49.253004 sh[651]: Success Mar 6 02:33:49.323267 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 6 02:33:49.323416 kernel: device-mapper: uevent: version 1.0.3 Mar 6 02:33:49.330596 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Mar 6 02:33:49.381153 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Mar 6 02:33:49.496483 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 6 02:33:49.531719 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 6 02:33:49.547287 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 6 02:33:49.603654 kernel: BTRFS: device fsid 1235dd15-5252-4928-9c6c-372370c6bfca devid 1 transid 39 /dev/mapper/usr (253:0) scanned by mount (663) Mar 6 02:33:49.603688 kernel: BTRFS info (device dm-0): first mount of filesystem 1235dd15-5252-4928-9c6c-372370c6bfca Mar 6 02:33:49.603699 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 6 02:33:49.651585 kernel: BTRFS info (device dm-0 state E): disabling log replay at mount time Mar 6 02:33:49.651677 kernel: BTRFS info (device dm-0 state E): enabling free space tree Mar 6 02:33:49.655275 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 6 02:33:49.667537 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Mar 6 02:33:49.676914 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 6 02:33:49.678698 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 6 02:33:49.686950 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 6 02:33:49.783097 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (696) Mar 6 02:33:49.791944 kernel: BTRFS info (device vda6): first mount of filesystem 993ea71e-e97d-4f5e-b5c7-fdac31a53b6b Mar 6 02:33:49.792002 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 6 02:33:49.828911 kernel: BTRFS info (device vda6): turning on async discard Mar 6 02:33:49.828991 kernel: BTRFS info (device vda6): enabling free space tree Mar 6 02:33:49.850085 kernel: BTRFS info (device vda6): last unmount of filesystem 993ea71e-e97d-4f5e-b5c7-fdac31a53b6b Mar 6 02:33:49.857196 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 6 02:33:49.873648 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 6 02:33:50.141225 ignition[751]: Ignition 2.22.0 Mar 6 02:33:50.142386 ignition[751]: Stage: fetch-offline Mar 6 02:33:50.144524 ignition[751]: no configs at "/usr/lib/ignition/base.d" Mar 6 02:33:50.144540 ignition[751]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 02:33:50.144649 ignition[751]: parsed url from cmdline: "" Mar 6 02:33:50.144654 ignition[751]: no config URL provided Mar 6 02:33:50.144661 ignition[751]: reading system config file "/usr/lib/ignition/user.ign" Mar 6 02:33:50.144676 ignition[751]: no config at "/usr/lib/ignition/user.ign" Mar 6 02:33:50.144862 ignition[751]: op(1): [started] loading QEMU firmware config module Mar 6 02:33:50.144876 ignition[751]: op(1): executing: "modprobe" "qemu_fw_cfg" Mar 6 02:33:50.182642 ignition[751]: op(1): [finished] loading QEMU firmware config module Mar 6 02:33:50.232733 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 6 02:33:50.253528 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 6 02:33:50.392131 systemd-networkd[840]: lo: Link UP Mar 6 02:33:50.392200 systemd-networkd[840]: lo: Gained carrier Mar 6 02:33:50.395121 systemd-networkd[840]: Enumeration completed Mar 6 02:33:50.395944 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 6 02:33:50.401670 systemd-networkd[840]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 02:33:50.401676 systemd-networkd[840]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 6 02:33:50.403239 systemd-networkd[840]: eth0: Link UP Mar 6 02:33:50.410464 systemd-networkd[840]: eth0: Gained carrier Mar 6 02:33:50.410479 systemd-networkd[840]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 02:33:50.432158 systemd[1]: Reached target network.target - Network. Mar 6 02:33:50.496723 systemd-networkd[840]: eth0: DHCPv4 address 10.0.0.69/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 6 02:33:51.271963 ignition[751]: parsing config with SHA512: ffc817f0e5eba5eb24add21e0a18f334251b5fbe12705ed20a0c09267f932744e290b7aa94aeaf3f6b841241c63385e8a6a88d37a9f7660934d6cdfd0eb3d342 Mar 6 02:33:51.315001 unknown[751]: fetched base config from "system" Mar 6 02:33:51.315019 unknown[751]: fetched user config from "qemu" Mar 6 02:33:51.331593 ignition[751]: fetch-offline: fetch-offline passed Mar 6 02:33:51.340143 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 6 02:33:51.331705 ignition[751]: Ignition finished successfully Mar 6 02:33:51.350498 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Mar 6 02:33:51.358622 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 6 02:33:51.600945 ignition[845]: Ignition 2.22.0 Mar 6 02:33:51.600959 ignition[845]: Stage: kargs Mar 6 02:33:51.601141 ignition[845]: no configs at "/usr/lib/ignition/base.d" Mar 6 02:33:51.601157 ignition[845]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 02:33:51.604415 ignition[845]: kargs: kargs passed Mar 6 02:33:51.605286 ignition[845]: Ignition finished successfully Mar 6 02:33:51.678624 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 6 02:33:51.692460 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 6 02:33:51.814463 ignition[853]: Ignition 2.22.0 Mar 6 02:33:51.814549 ignition[853]: Stage: disks Mar 6 02:33:51.814931 ignition[853]: no configs at "/usr/lib/ignition/base.d" Mar 6 02:33:51.855472 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 6 02:33:51.814949 ignition[853]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 02:33:51.820202 ignition[853]: disks: disks passed Mar 6 02:33:51.890264 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 6 02:33:51.820266 ignition[853]: Ignition finished successfully Mar 6 02:33:51.900684 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 6 02:33:51.941960 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 6 02:33:51.957661 systemd[1]: Reached target sysinit.target - System Initialization. Mar 6 02:33:51.964017 systemd[1]: Reached target basic.target - Basic System. Mar 6 02:33:51.985189 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 6 02:33:52.199049 systemd-fsck[863]: ROOT: clean, 15/553520 files, 52789/553472 blocks Mar 6 02:33:52.223491 systemd-networkd[840]: eth0: Gained IPv6LL Mar 6 02:33:52.239530 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 6 02:33:52.276257 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 6 02:33:54.985455 kernel: EXT4-fs (vda9): mounted filesystem 16ab7223-a8af-43d2-ad40-7e1bf0ff2a89 r/w with ordered data mode. Quota mode: none. Mar 6 02:33:55.000674 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 6 02:33:55.068436 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 6 02:33:55.141685 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 6 02:33:55.176010 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 6 02:33:55.207433 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 6 02:33:55.207707 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 6 02:33:55.207907 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 6 02:33:55.395417 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (871) Mar 6 02:33:57.156011 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 1507609765 wd_nsec: 1507609144 Mar 6 02:33:57.158387 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 6 02:33:57.243897 kernel: BTRFS info (device vda6): first mount of filesystem 993ea71e-e97d-4f5e-b5c7-fdac31a53b6b Mar 6 02:33:57.244017 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 6 02:33:57.243643 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 6 02:33:57.368133 kernel: BTRFS info (device vda6): turning on async discard Mar 6 02:33:57.368599 kernel: BTRFS info (device vda6): enabling free space tree Mar 6 02:33:57.387475 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 6 02:33:57.954597 initrd-setup-root[895]: cut: /sysroot/etc/passwd: No such file or directory Mar 6 02:33:57.991559 initrd-setup-root[902]: cut: /sysroot/etc/group: No such file or directory Mar 6 02:33:58.060452 initrd-setup-root[909]: cut: /sysroot/etc/shadow: No such file or directory Mar 6 02:33:58.147243 initrd-setup-root[916]: cut: /sysroot/etc/gshadow: No such file or directory Mar 6 02:33:58.999593 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 6 02:33:59.054031 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 6 02:33:59.097603 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 6 02:33:59.156401 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 6 02:33:59.192188 kernel: BTRFS info (device vda6): last unmount of filesystem 993ea71e-e97d-4f5e-b5c7-fdac31a53b6b Mar 6 02:33:59.313999 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 6 02:33:59.497406 ignition[984]: INFO : Ignition 2.22.0 Mar 6 02:33:59.497406 ignition[984]: INFO : Stage: mount Mar 6 02:33:59.497406 ignition[984]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 6 02:33:59.497406 ignition[984]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 02:33:59.595223 ignition[984]: INFO : mount: mount passed Mar 6 02:33:59.595223 ignition[984]: INFO : Ignition finished successfully Mar 6 02:33:59.547182 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 6 02:33:59.587253 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 6 02:33:59.716494 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 6 02:33:59.850495 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (997) Mar 6 02:33:59.867025 kernel: BTRFS info (device vda6): first mount of filesystem 993ea71e-e97d-4f5e-b5c7-fdac31a53b6b Mar 6 02:33:59.867536 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 6 02:33:59.941215 kernel: BTRFS info (device vda6): turning on async discard Mar 6 02:33:59.942829 kernel: BTRFS info (device vda6): enabling free space tree Mar 6 02:33:59.957684 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 6 02:34:00.271181 ignition[1014]: INFO : Ignition 2.22.0 Mar 6 02:34:00.271181 ignition[1014]: INFO : Stage: files Mar 6 02:34:00.271181 ignition[1014]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 6 02:34:00.271181 ignition[1014]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 02:34:00.336256 ignition[1014]: DEBUG : files: compiled without relabeling support, skipping Mar 6 02:34:00.336256 ignition[1014]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 6 02:34:00.336256 ignition[1014]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 6 02:34:00.336256 ignition[1014]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 6 02:34:00.336256 ignition[1014]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 6 02:34:00.435181 ignition[1014]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 6 02:34:00.435181 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 6 02:34:00.435181 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Mar 6 02:34:00.343547 unknown[1014]: wrote ssh authorized keys file for user: core Mar 6 02:34:00.609733 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 6 02:34:01.348554 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 6 02:34:01.348554 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Mar 6 02:34:01.392056 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Mar 6 02:34:01.769580 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.35.1-x86-64.raw: attempt #1 Mar 6 02:34:02.473213 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Mar 6 02:34:11.608939 ignition[1014]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Mar 6 02:34:11.608939 ignition[1014]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Mar 6 02:34:11.660273 ignition[1014]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 6 02:34:12.573957 ignition[1014]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 6 02:34:12.573957 ignition[1014]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Mar 6 02:34:12.573957 ignition[1014]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Mar 6 02:34:12.645101 ignition[1014]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 6 02:34:12.645101 ignition[1014]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 6 02:34:12.645101 ignition[1014]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Mar 6 02:34:12.645101 ignition[1014]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Mar 6 02:34:12.982593 ignition[1014]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Mar 6 02:34:13.190915 ignition[1014]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Mar 6 02:34:13.208125 ignition[1014]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Mar 6 02:34:13.208125 ignition[1014]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Mar 6 02:34:13.208125 ignition[1014]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Mar 6 02:34:13.208125 ignition[1014]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 6 02:34:13.208125 ignition[1014]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 6 02:34:13.208125 ignition[1014]: INFO : files: files passed Mar 6 02:34:13.208125 ignition[1014]: INFO : Ignition finished successfully Mar 6 02:34:13.294261 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 6 02:34:13.397177 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 6 02:34:13.420211 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 6 02:34:13.546885 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 6 02:34:13.583994 initrd-setup-root-after-ignition[1042]: grep: /sysroot/oem/oem-release: No such file or directory Mar 6 02:34:13.547185 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 6 02:34:13.638611 initrd-setup-root-after-ignition[1049]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 6 02:34:13.621063 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 6 02:34:13.689164 initrd-setup-root-after-ignition[1045]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 6 02:34:13.689164 initrd-setup-root-after-ignition[1045]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 6 02:34:13.656547 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 6 02:34:13.708199 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 6 02:34:14.215069 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 6 02:34:14.215454 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 6 02:34:14.252162 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 6 02:34:14.283546 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 6 02:34:14.364000 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 6 02:34:14.372097 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 6 02:34:14.595304 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 6 02:34:14.647034 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 6 02:34:14.788918 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 6 02:34:14.822293 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 6 02:34:14.850288 systemd[1]: Stopped target timers.target - Timer Units. Mar 6 02:34:14.901527 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 6 02:34:14.902161 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 6 02:34:14.964121 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 6 02:34:14.983991 systemd[1]: Stopped target basic.target - Basic System. Mar 6 02:34:15.003561 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 6 02:34:15.062500 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 6 02:34:15.081561 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 6 02:34:15.099233 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Mar 6 02:34:15.116703 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 6 02:34:15.142193 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 6 02:34:15.168070 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 6 02:34:15.169022 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 6 02:34:15.253240 systemd[1]: Stopped target swap.target - Swaps. Mar 6 02:34:15.298232 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 6 02:34:15.356187 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 6 02:34:15.398291 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 6 02:34:15.443526 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 6 02:34:15.444212 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 6 02:34:15.490984 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 6 02:34:15.492191 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 6 02:34:15.492567 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 6 02:34:15.600164 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 6 02:34:15.602184 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 6 02:34:15.659156 systemd[1]: Stopped target paths.target - Path Units. Mar 6 02:34:15.675048 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 6 02:34:15.677051 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 6 02:34:15.699428 systemd[1]: Stopped target slices.target - Slice Units. Mar 6 02:34:15.712000 systemd[1]: Stopped target sockets.target - Socket Units. Mar 6 02:34:15.765622 systemd[1]: iscsid.socket: Deactivated successfully. Mar 6 02:34:15.766261 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 6 02:34:15.805283 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 6 02:34:15.805518 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 6 02:34:15.844068 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 6 02:34:15.844901 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 6 02:34:15.890931 systemd[1]: ignition-files.service: Deactivated successfully. Mar 6 02:34:15.891108 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 6 02:34:15.906114 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 6 02:34:15.933447 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 6 02:34:15.934105 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 6 02:34:15.978310 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 6 02:34:16.001527 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 6 02:34:16.002239 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 6 02:34:16.045681 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 6 02:34:16.045995 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 6 02:34:16.083961 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 6 02:34:16.084146 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 6 02:34:16.366538 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 6 02:34:16.421154 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 6 02:34:16.431653 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 6 02:34:16.495661 ignition[1069]: INFO : Ignition 2.22.0 Mar 6 02:34:16.495661 ignition[1069]: INFO : Stage: umount Mar 6 02:34:16.538125 ignition[1069]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 6 02:34:16.538125 ignition[1069]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 02:34:16.538125 ignition[1069]: INFO : umount: umount passed Mar 6 02:34:16.538125 ignition[1069]: INFO : Ignition finished successfully Mar 6 02:34:16.636555 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 6 02:34:16.643123 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 6 02:34:16.707428 systemd[1]: Stopped target network.target - Network. Mar 6 02:34:16.774078 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 6 02:34:16.778265 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 6 02:34:16.829265 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 6 02:34:16.836138 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 6 02:34:16.867668 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 6 02:34:16.868166 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 6 02:34:16.939617 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 6 02:34:16.940014 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 6 02:34:16.999563 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 6 02:34:17.006240 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 6 02:34:17.080189 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 6 02:34:17.248693 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 6 02:34:17.374048 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 6 02:34:17.376275 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 6 02:34:17.464559 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Mar 6 02:34:17.466729 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 6 02:34:17.467502 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 6 02:34:17.541245 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Mar 6 02:34:17.548688 systemd[1]: Stopped target network-pre.target - Preparation for Network. Mar 6 02:34:17.577114 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 6 02:34:17.577180 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 6 02:34:17.630270 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 6 02:34:17.645649 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 6 02:34:17.646052 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 6 02:34:17.663458 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 6 02:34:17.663558 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 6 02:34:17.694053 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 6 02:34:17.694128 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 6 02:34:17.739163 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 6 02:34:17.740497 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 6 02:34:17.814672 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 6 02:34:17.875200 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 6 02:34:17.876129 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Mar 6 02:34:17.931434 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 6 02:34:17.931994 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 6 02:34:18.135452 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 6 02:34:18.136693 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 6 02:34:18.259466 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 6 02:34:18.260300 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 6 02:34:18.300670 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 6 02:34:18.301156 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 6 02:34:18.347005 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 6 02:34:18.349207 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 6 02:34:18.413466 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 6 02:34:18.413676 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 6 02:34:18.507582 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 6 02:34:18.566531 systemd[1]: systemd-network-generator.service: Deactivated successfully. Mar 6 02:34:18.567096 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Mar 6 02:34:18.687732 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 6 02:34:18.688200 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 6 02:34:18.776524 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 6 02:34:18.776707 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 02:34:18.837486 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Mar 6 02:34:18.837581 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Mar 6 02:34:18.837647 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 6 02:34:18.839631 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 6 02:34:18.840051 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 6 02:34:19.044144 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 6 02:34:19.046963 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 6 02:34:19.076015 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 6 02:34:19.133471 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 6 02:34:19.247653 systemd[1]: Switching root. Mar 6 02:34:19.351065 systemd-journald[204]: Journal stopped Mar 6 02:34:26.510282 systemd-journald[204]: Received SIGTERM from PID 1 (systemd). Mar 6 02:34:26.511074 kernel: SELinux: policy capability network_peer_controls=1 Mar 6 02:34:26.512449 kernel: SELinux: policy capability open_perms=1 Mar 6 02:34:26.512479 kernel: SELinux: policy capability extended_socket_class=1 Mar 6 02:34:26.512503 kernel: SELinux: policy capability always_check_network=0 Mar 6 02:34:26.512520 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 6 02:34:26.512541 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 6 02:34:26.512555 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 6 02:34:26.512569 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 6 02:34:26.512585 kernel: SELinux: policy capability userspace_initial_context=0 Mar 6 02:34:26.512601 kernel: audit: type=1403 audit(1772764460.080:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 6 02:34:26.512620 systemd[1]: Successfully loaded SELinux policy in 239.966ms. Mar 6 02:34:26.512649 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 33.154ms. Mar 6 02:34:26.512943 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 6 02:34:26.512963 systemd[1]: Detected virtualization kvm. Mar 6 02:34:26.512980 systemd[1]: Detected architecture x86-64. Mar 6 02:34:26.512999 systemd[1]: Detected first boot. Mar 6 02:34:26.513015 systemd[1]: Initializing machine ID from VM UUID. Mar 6 02:34:26.513030 zram_generator::config[1116]: No configuration found. Mar 6 02:34:26.513046 kernel: Guest personality initialized and is inactive Mar 6 02:34:26.513069 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Mar 6 02:34:26.513086 kernel: Initialized host personality Mar 6 02:34:26.513101 kernel: NET: Registered PF_VSOCK protocol family Mar 6 02:34:26.513115 systemd[1]: Populated /etc with preset unit settings. Mar 6 02:34:26.513132 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Mar 6 02:34:26.513148 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 6 02:34:26.513166 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 6 02:34:26.513183 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 6 02:34:26.513199 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 6 02:34:26.513220 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 6 02:34:26.513235 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 6 02:34:26.513251 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 6 02:34:26.513277 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 6 02:34:26.513294 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 6 02:34:26.513311 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 6 02:34:26.515947 systemd[1]: Created slice user.slice - User and Session Slice. Mar 6 02:34:26.516057 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 6 02:34:26.516087 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 6 02:34:26.516112 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 6 02:34:26.516128 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 6 02:34:26.516148 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 6 02:34:26.516164 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 6 02:34:26.516179 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 6 02:34:26.516195 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 6 02:34:26.516210 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 6 02:34:26.516232 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 6 02:34:26.516247 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 6 02:34:26.516262 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 6 02:34:26.516562 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 6 02:34:26.516579 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 6 02:34:26.516594 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 6 02:34:26.516609 systemd[1]: Reached target slices.target - Slice Units. Mar 6 02:34:26.516624 systemd[1]: Reached target swap.target - Swaps. Mar 6 02:34:26.516638 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 6 02:34:26.516659 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 6 02:34:26.516676 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Mar 6 02:34:26.516695 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 6 02:34:26.516713 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 6 02:34:26.516728 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 6 02:34:26.517054 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 6 02:34:26.517075 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 6 02:34:26.517092 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 6 02:34:26.517108 systemd[1]: Mounting media.mount - External Media Directory... Mar 6 02:34:26.517130 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 02:34:26.517149 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 6 02:34:26.517167 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 6 02:34:26.517182 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 6 02:34:26.517199 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 6 02:34:26.517215 systemd[1]: Reached target machines.target - Containers. Mar 6 02:34:26.517542 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 6 02:34:26.517560 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 6 02:34:26.517580 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 6 02:34:26.517595 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 6 02:34:26.517611 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 6 02:34:26.517627 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 6 02:34:26.517642 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 6 02:34:26.517659 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 6 02:34:26.517674 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 6 02:34:26.517689 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 6 02:34:26.517703 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 6 02:34:26.517722 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 6 02:34:26.518061 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 6 02:34:26.518087 systemd[1]: Stopped systemd-fsck-usr.service. Mar 6 02:34:26.518107 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 6 02:34:26.518124 kernel: ACPI: bus type drm_connector registered Mar 6 02:34:26.518139 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 6 02:34:26.518153 kernel: fuse: init (API version 7.41) Mar 6 02:34:26.518168 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 6 02:34:26.518183 kernel: loop: module loaded Mar 6 02:34:26.518205 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 6 02:34:26.518524 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 6 02:34:26.518590 systemd-journald[1201]: Collecting audit messages is disabled. Mar 6 02:34:26.518632 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Mar 6 02:34:26.518649 systemd-journald[1201]: Journal started Mar 6 02:34:26.518678 systemd-journald[1201]: Runtime Journal (/run/log/journal/86b8fd27b29f4970962f2d2e02c7221e) is 6M, max 48.1M, 42.1M free. Mar 6 02:34:23.870442 systemd[1]: Queued start job for default target multi-user.target. Mar 6 02:34:23.898580 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 6 02:34:23.902920 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 6 02:34:23.903919 systemd[1]: systemd-journald.service: Consumed 2.730s CPU time. Mar 6 02:34:26.605031 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 6 02:34:26.639082 systemd[1]: verity-setup.service: Deactivated successfully. Mar 6 02:34:26.641145 systemd[1]: Stopped verity-setup.service. Mar 6 02:34:26.692595 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 02:34:26.758282 systemd[1]: Started systemd-journald.service - Journal Service. Mar 6 02:34:26.792175 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 6 02:34:26.809515 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 6 02:34:26.843513 systemd[1]: Mounted media.mount - External Media Directory. Mar 6 02:34:26.858552 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 6 02:34:26.874535 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 6 02:34:26.890205 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 6 02:34:26.932654 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 6 02:34:26.953152 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 6 02:34:26.971032 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 6 02:34:26.972178 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 6 02:34:26.989712 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 6 02:34:26.990909 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 6 02:34:27.013973 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 6 02:34:27.015050 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 6 02:34:27.053269 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 6 02:34:27.054960 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 6 02:34:27.087176 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 6 02:34:27.088629 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 6 02:34:27.131300 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 6 02:34:27.132277 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 6 02:34:27.160528 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 6 02:34:27.178044 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 6 02:34:27.230577 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 6 02:34:27.282082 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Mar 6 02:34:27.307541 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 6 02:34:27.375309 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 6 02:34:27.405017 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 6 02:34:27.427641 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 6 02:34:27.439964 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 6 02:34:27.440101 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 6 02:34:27.457566 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Mar 6 02:34:27.482237 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 6 02:34:27.496073 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 6 02:34:27.500486 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 6 02:34:27.521956 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 6 02:34:27.545685 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 6 02:34:27.569088 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 6 02:34:27.602578 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 6 02:34:27.653650 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 6 02:34:27.743581 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 6 02:34:27.754023 systemd-journald[1201]: Time spent on flushing to /var/log/journal/86b8fd27b29f4970962f2d2e02c7221e is 64.216ms for 1066 entries. Mar 6 02:34:27.754023 systemd-journald[1201]: System Journal (/var/log/journal/86b8fd27b29f4970962f2d2e02c7221e) is 8M, max 195.6M, 187.6M free. Mar 6 02:34:27.903997 systemd-journald[1201]: Received client request to flush runtime journal. Mar 6 02:34:27.820534 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 6 02:34:27.844640 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 6 02:34:27.875181 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 6 02:34:27.937509 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 6 02:34:27.979944 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 6 02:34:28.005283 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 6 02:34:28.060939 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Mar 6 02:34:28.171266 kernel: loop0: detected capacity change from 0 to 110984 Mar 6 02:34:28.230265 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 6 02:34:28.362154 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 6 02:34:28.499432 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 6 02:34:28.506722 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 6 02:34:28.536460 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Mar 6 02:34:28.585507 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 6 02:34:28.611092 kernel: loop1: detected capacity change from 0 to 217752 Mar 6 02:34:28.937032 kernel: loop2: detected capacity change from 0 to 128560 Mar 6 02:34:29.591552 systemd-tmpfiles[1253]: ACLs are not supported, ignoring. Mar 6 02:34:29.591576 systemd-tmpfiles[1253]: ACLs are not supported, ignoring. Mar 6 02:34:29.705735 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 6 02:34:30.129598 kernel: loop3: detected capacity change from 0 to 110984 Mar 6 02:34:30.435937 kernel: loop4: detected capacity change from 0 to 217752 Mar 6 02:34:30.612992 kernel: loop5: detected capacity change from 0 to 128560 Mar 6 02:34:30.943050 (sd-merge)[1259]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Mar 6 02:34:30.947113 (sd-merge)[1259]: Merged extensions into '/usr'. Mar 6 02:34:31.062970 systemd[1]: Reload requested from client PID 1236 ('systemd-sysext') (unit systemd-sysext.service)... Mar 6 02:34:31.063001 systemd[1]: Reloading... Mar 6 02:34:31.566948 zram_generator::config[1286]: No configuration found. Mar 6 02:34:32.843691 systemd[1]: Reloading finished in 1778 ms. Mar 6 02:34:33.113586 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 6 02:34:33.207901 systemd[1]: Starting ensure-sysext.service... Mar 6 02:34:33.254238 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 6 02:34:33.490998 ldconfig[1231]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 6 02:34:33.535653 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 6 02:34:33.557444 systemd[1]: Reload requested from client PID 1322 ('systemctl') (unit ensure-sysext.service)... Mar 6 02:34:33.557462 systemd[1]: Reloading... Mar 6 02:34:33.748235 systemd-tmpfiles[1323]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Mar 6 02:34:33.748455 systemd-tmpfiles[1323]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Mar 6 02:34:33.750151 systemd-tmpfiles[1323]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 6 02:34:33.751020 systemd-tmpfiles[1323]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 6 02:34:33.753280 systemd-tmpfiles[1323]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 6 02:34:33.754106 systemd-tmpfiles[1323]: ACLs are not supported, ignoring. Mar 6 02:34:33.754204 systemd-tmpfiles[1323]: ACLs are not supported, ignoring. Mar 6 02:34:33.767142 systemd-tmpfiles[1323]: Detected autofs mount point /boot during canonicalization of boot. Mar 6 02:34:33.767166 systemd-tmpfiles[1323]: Skipping /boot Mar 6 02:34:33.801278 systemd-tmpfiles[1323]: Detected autofs mount point /boot during canonicalization of boot. Mar 6 02:34:33.801301 systemd-tmpfiles[1323]: Skipping /boot Mar 6 02:34:33.944946 zram_generator::config[1348]: No configuration found. Mar 6 02:34:34.683416 systemd[1]: Reloading finished in 1125 ms. Mar 6 02:34:34.851558 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 6 02:34:35.083579 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 6 02:34:35.181542 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 6 02:34:35.288455 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 6 02:34:35.384275 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 6 02:34:35.452446 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 6 02:34:35.509146 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 6 02:34:35.598266 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 6 02:34:35.653966 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 02:34:35.658500 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 6 02:34:35.663108 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 6 02:34:35.781999 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 6 02:34:35.976415 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 6 02:34:35.990009 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 6 02:34:35.990262 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 6 02:34:36.003633 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 6 02:34:36.029185 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 02:34:36.037599 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 6 02:34:36.057310 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 6 02:34:36.058281 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 6 02:34:36.083464 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 6 02:34:36.083924 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 6 02:34:36.108712 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 6 02:34:36.109295 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 6 02:34:36.133592 augenrules[1419]: No rules Mar 6 02:34:36.139047 systemd-udevd[1402]: Using default interface naming scheme 'v255'. Mar 6 02:34:36.142221 systemd[1]: audit-rules.service: Deactivated successfully. Mar 6 02:34:36.143554 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 6 02:34:36.215159 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 6 02:34:36.248560 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 6 02:34:36.283904 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 02:34:36.288297 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 6 02:34:36.304571 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 6 02:34:36.314514 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 6 02:34:36.350184 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 6 02:34:36.373163 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 6 02:34:36.392537 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 6 02:34:36.403991 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 6 02:34:36.404175 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 6 02:34:36.410101 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 6 02:34:36.427212 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 6 02:34:36.427517 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 02:34:36.430533 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 6 02:34:36.446130 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 6 02:34:36.465926 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 6 02:34:36.466231 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 6 02:34:36.520583 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 6 02:34:36.523264 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 6 02:34:36.532554 augenrules[1431]: /sbin/augenrules: No change Mar 6 02:34:36.537051 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 6 02:34:36.544162 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 6 02:34:36.565511 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 6 02:34:36.598558 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 6 02:34:36.599136 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 6 02:34:36.621573 augenrules[1477]: No rules Mar 6 02:34:36.623505 systemd[1]: audit-rules.service: Deactivated successfully. Mar 6 02:34:36.625263 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 6 02:34:36.674458 systemd[1]: Finished ensure-sysext.service. Mar 6 02:34:36.763231 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 6 02:34:36.789171 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 6 02:34:36.789428 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 6 02:34:36.800110 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 6 02:34:36.913918 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Mar 6 02:34:37.257036 systemd-resolved[1399]: Positive Trust Anchors: Mar 6 02:34:37.257245 systemd-resolved[1399]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 6 02:34:37.257285 systemd-resolved[1399]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 6 02:34:37.328601 systemd-resolved[1399]: Defaulting to hostname 'linux'. Mar 6 02:34:37.336605 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 6 02:34:37.366032 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 6 02:34:37.593225 kernel: mousedev: PS/2 mouse device common for all mice Mar 6 02:34:37.731209 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 6 02:34:37.745172 systemd[1]: Reached target sysinit.target - System Initialization. Mar 6 02:34:37.756963 systemd-networkd[1497]: lo: Link UP Mar 6 02:34:37.756971 systemd-networkd[1497]: lo: Gained carrier Mar 6 02:34:37.762224 systemd-networkd[1497]: Enumeration completed Mar 6 02:34:37.768977 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 6 02:34:37.768997 systemd-networkd[1497]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 02:34:37.769003 systemd-networkd[1497]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 6 02:34:37.778585 systemd-networkd[1497]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 02:34:37.778649 systemd-networkd[1497]: eth0: Link UP Mar 6 02:34:37.780110 systemd-networkd[1497]: eth0: Gained carrier Mar 6 02:34:37.780133 systemd-networkd[1497]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 02:34:37.788661 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 6 02:34:37.821404 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input4 Mar 6 02:34:37.825074 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Mar 6 02:34:37.848569 kernel: ACPI: button: Power Button [PWRF] Mar 6 02:34:37.872060 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 6 02:34:37.902529 systemd-networkd[1497]: eth0: DHCPv4 address 10.0.0.69/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 6 02:34:37.906237 systemd-timesyncd[1498]: Network configuration changed, trying to establish connection. Mar 6 02:34:37.907116 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 6 02:34:37.907167 systemd[1]: Reached target paths.target - Path Units. Mar 6 02:34:37.936543 systemd[1]: Reached target time-set.target - System Time Set. Mar 6 02:34:37.969489 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 6 02:34:39.396254 systemd-timesyncd[1498]: Contacted time server 10.0.0.1:123 (10.0.0.1). Mar 6 02:34:39.396757 systemd-timesyncd[1498]: Initial clock synchronization to Fri 2026-03-06 02:34:39.388499 UTC. Mar 6 02:34:39.397374 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 6 02:34:39.400520 systemd-resolved[1399]: Clock change detected. Flushing caches. Mar 6 02:34:39.465639 systemd[1]: Reached target timers.target - Timer Units. Mar 6 02:34:39.527727 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 6 02:34:39.551618 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 6 02:34:39.567631 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Mar 6 02:34:39.582062 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Mar 6 02:34:39.602116 systemd[1]: Reached target ssh-access.target - SSH Access Available. Mar 6 02:34:39.717605 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 6 02:34:39.742487 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Mar 6 02:34:39.778680 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 6 02:34:39.847030 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 6 02:34:39.896606 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 6 02:34:39.933025 systemd[1]: Reached target network.target - Network. Mar 6 02:34:39.970763 systemd[1]: Reached target sockets.target - Socket Units. Mar 6 02:34:40.020170 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Mar 6 02:34:40.053021 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 6 02:34:40.053705 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 6 02:34:40.055214 systemd[1]: Reached target basic.target - Basic System. Mar 6 02:34:40.094269 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 6 02:34:40.094775 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 6 02:34:40.106740 systemd[1]: Starting containerd.service - containerd container runtime... Mar 6 02:34:40.150225 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 6 02:34:40.190065 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 6 02:34:40.264718 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 6 02:34:40.290088 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 6 02:34:40.334574 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 6 02:34:40.395537 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Mar 6 02:34:40.519205 jq[1532]: false Mar 6 02:34:40.520238 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 6 02:34:40.564970 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 6 02:34:40.589963 systemd-networkd[1497]: eth0: Gained IPv6LL Mar 6 02:34:40.591645 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 6 02:34:40.646611 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 6 02:34:40.662509 extend-filesystems[1533]: Found /dev/vda6 Mar 6 02:34:40.682156 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 6 02:34:40.862674 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 6 02:34:41.021734 extend-filesystems[1533]: Found /dev/vda9 Mar 6 02:34:41.011222 oslogin_cache_refresh[1534]: Refreshing passwd entry cache Mar 6 02:34:41.079001 google_oslogin_nss_cache[1534]: oslogin_cache_refresh[1534]: Refreshing passwd entry cache Mar 6 02:34:40.971051 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Mar 6 02:34:41.041225 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 6 02:34:41.073126 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 6 02:34:41.075162 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 6 02:34:41.087532 extend-filesystems[1533]: Checking size of /dev/vda9 Mar 6 02:34:41.103701 google_oslogin_nss_cache[1534]: oslogin_cache_refresh[1534]: Failure getting users, quitting Mar 6 02:34:41.103701 google_oslogin_nss_cache[1534]: oslogin_cache_refresh[1534]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Mar 6 02:34:41.103701 google_oslogin_nss_cache[1534]: oslogin_cache_refresh[1534]: Refreshing group entry cache Mar 6 02:34:41.089645 oslogin_cache_refresh[1534]: Failure getting users, quitting Mar 6 02:34:41.089759 oslogin_cache_refresh[1534]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Mar 6 02:34:41.089826 oslogin_cache_refresh[1534]: Refreshing group entry cache Mar 6 02:34:41.105265 systemd[1]: Starting update-engine.service - Update Engine... Mar 6 02:34:41.122169 oslogin_cache_refresh[1534]: Failure getting groups, quitting Mar 6 02:34:41.131958 google_oslogin_nss_cache[1534]: oslogin_cache_refresh[1534]: Failure getting groups, quitting Mar 6 02:34:41.131958 google_oslogin_nss_cache[1534]: oslogin_cache_refresh[1534]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Mar 6 02:34:41.122195 oslogin_cache_refresh[1534]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Mar 6 02:34:41.135812 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 6 02:34:41.159464 extend-filesystems[1533]: Resized partition /dev/vda9 Mar 6 02:34:41.204970 extend-filesystems[1558]: resize2fs 1.47.3 (8-Jul-2025) Mar 6 02:34:41.241584 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Mar 6 02:34:41.260827 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 6 02:34:41.263053 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 6 02:34:41.263676 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 6 02:34:41.265551 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Mar 6 02:34:41.266089 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Mar 6 02:34:41.291038 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 6 02:34:41.293499 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 6 02:34:41.324853 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 6 02:34:41.383023 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 6 02:34:41.390594 jq[1555]: true Mar 6 02:34:41.455635 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Mar 6 02:34:41.513071 systemd[1]: motdgen.service: Deactivated successfully. Mar 6 02:34:41.515108 extend-filesystems[1558]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 6 02:34:41.515108 extend-filesystems[1558]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 6 02:34:41.515108 extend-filesystems[1558]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Mar 6 02:34:41.613685 extend-filesystems[1533]: Resized filesystem in /dev/vda9 Mar 6 02:34:41.516013 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 6 02:34:41.561746 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 6 02:34:41.600817 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 6 02:34:41.684123 update_engine[1551]: I20260306 02:34:41.675019 1551 main.cc:92] Flatcar Update Engine starting Mar 6 02:34:41.688209 jq[1568]: true Mar 6 02:34:41.696198 systemd-logind[1544]: Watching system buttons on /dev/input/event2 (Power Button) Mar 6 02:34:41.696559 systemd-logind[1544]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 6 02:34:41.699716 systemd-logind[1544]: New seat seat0. Mar 6 02:34:41.709137 (ntainerd)[1569]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 6 02:34:41.711486 systemd[1]: Started systemd-logind.service - User Login Management. Mar 6 02:34:41.809168 tar[1562]: linux-amd64/LICENSE Mar 6 02:34:41.826144 tar[1562]: linux-amd64/helm Mar 6 02:34:41.873611 dbus-daemon[1530]: [system] SELinux support is enabled Mar 6 02:34:41.882253 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 6 02:34:41.917064 update_engine[1551]: I20260306 02:34:41.916092 1551 update_check_scheduler.cc:74] Next update check in 2m59s Mar 6 02:34:41.960840 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Mar 6 02:34:41.993085 dbus-daemon[1530]: [system] Successfully activated service 'org.freedesktop.systemd1' Mar 6 02:34:42.068671 systemd[1]: Started update-engine.service - Update Engine. Mar 6 02:34:42.086116 systemd[1]: Reached target network-online.target - Network is Online. Mar 6 02:34:42.108486 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Mar 6 02:34:42.154829 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:34:42.208755 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 6 02:34:42.222260 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 6 02:34:42.223577 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 6 02:34:42.247523 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 02:34:42.261623 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 6 02:34:42.261815 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 6 02:34:42.329772 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 6 02:34:42.545199 bash[1604]: Updated "/home/core/.ssh/authorized_keys" Mar 6 02:34:42.634682 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 6 02:34:42.666016 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 6 02:34:42.951102 systemd[1]: coreos-metadata.service: Deactivated successfully. Mar 6 02:34:42.951974 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Mar 6 02:34:42.959086 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 6 02:34:42.985706 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 6 02:34:43.126273 sshd_keygen[1550]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 6 02:34:43.789259 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 6 02:34:43.834087 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 6 02:34:44.032073 systemd[1]: issuegen.service: Deactivated successfully. Mar 6 02:34:44.032960 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 6 02:34:44.182046 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 6 02:34:44.396487 locksmithd[1606]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 6 02:34:44.567596 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 02:34:44.768677 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 6 02:34:44.927239 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 6 02:34:44.954997 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 6 02:34:44.972985 systemd[1]: Reached target getty.target - Login Prompts. Mar 6 02:34:46.120567 kernel: kvm_amd: TSC scaling supported Mar 6 02:34:46.121632 kernel: kvm_amd: Nested Virtualization enabled Mar 6 02:34:46.121675 kernel: kvm_amd: Nested Paging enabled Mar 6 02:34:46.144690 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Mar 6 02:34:46.144776 kernel: kvm_amd: PMU virtualization is disabled Mar 6 02:34:46.412078 containerd[1569]: time="2026-03-06T02:34:46Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Mar 6 02:34:46.419595 containerd[1569]: time="2026-03-06T02:34:46.419065138Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Mar 6 02:34:46.496480 containerd[1569]: time="2026-03-06T02:34:46.491696734Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="13.675µs" Mar 6 02:34:46.496480 containerd[1569]: time="2026-03-06T02:34:46.491754291Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Mar 6 02:34:46.496480 containerd[1569]: time="2026-03-06T02:34:46.491789567Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Mar 6 02:34:46.496480 containerd[1569]: time="2026-03-06T02:34:46.493019073Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Mar 6 02:34:46.496480 containerd[1569]: time="2026-03-06T02:34:46.493053176Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Mar 6 02:34:46.496480 containerd[1569]: time="2026-03-06T02:34:46.493098521Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 6 02:34:46.496480 containerd[1569]: time="2026-03-06T02:34:46.493195382Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 6 02:34:46.496480 containerd[1569]: time="2026-03-06T02:34:46.493217162Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 6 02:34:46.516499 containerd[1569]: time="2026-03-06T02:34:46.516131552Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 6 02:34:46.516499 containerd[1569]: time="2026-03-06T02:34:46.516259290Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 6 02:34:46.516499 containerd[1569]: time="2026-03-06T02:34:46.516462590Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 6 02:34:46.516499 containerd[1569]: time="2026-03-06T02:34:46.516479512Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Mar 6 02:34:46.516685 containerd[1569]: time="2026-03-06T02:34:46.516633279Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Mar 6 02:34:46.518046 containerd[1569]: time="2026-03-06T02:34:46.517225274Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 6 02:34:46.518113 containerd[1569]: time="2026-03-06T02:34:46.518090610Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 6 02:34:46.518113 containerd[1569]: time="2026-03-06T02:34:46.518109124Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Mar 6 02:34:46.518721 containerd[1569]: time="2026-03-06T02:34:46.518504863Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Mar 6 02:34:46.520510 containerd[1569]: time="2026-03-06T02:34:46.519702070Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Mar 6 02:34:46.520510 containerd[1569]: time="2026-03-06T02:34:46.519971121Z" level=info msg="metadata content store policy set" policy=shared Mar 6 02:34:46.549061 containerd[1569]: time="2026-03-06T02:34:46.548761545Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549176741Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549451334Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549472783Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549489765Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549504793Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549520934Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549535110Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549548996Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549569524Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549581236Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Mar 6 02:34:46.549734 containerd[1569]: time="2026-03-06T02:34:46.549596424Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Mar 6 02:34:46.550237 containerd[1569]: time="2026-03-06T02:34:46.549951798Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Mar 6 02:34:46.550237 containerd[1569]: time="2026-03-06T02:34:46.549981955Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Mar 6 02:34:46.552738 containerd[1569]: time="2026-03-06T02:34:46.550267718Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Mar 6 02:34:46.552738 containerd[1569]: time="2026-03-06T02:34:46.552663582Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Mar 6 02:34:46.552738 containerd[1569]: time="2026-03-06T02:34:46.552689170Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Mar 6 02:34:46.552738 containerd[1569]: time="2026-03-06T02:34:46.552705470Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Mar 6 02:34:46.552738 containerd[1569]: time="2026-03-06T02:34:46.552723463Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Mar 6 02:34:46.552738 containerd[1569]: time="2026-03-06T02:34:46.552737409Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Mar 6 02:34:46.553057 containerd[1569]: time="2026-03-06T02:34:46.552753790Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Mar 6 02:34:46.553057 containerd[1569]: time="2026-03-06T02:34:46.552769790Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Mar 6 02:34:46.553057 containerd[1569]: time="2026-03-06T02:34:46.552784147Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Mar 6 02:34:46.553057 containerd[1569]: time="2026-03-06T02:34:46.552854609Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Mar 6 02:34:46.553057 containerd[1569]: time="2026-03-06T02:34:46.552977508Z" level=info msg="Start snapshots syncer" Mar 6 02:34:46.553187 containerd[1569]: time="2026-03-06T02:34:46.553120044Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Mar 6 02:34:46.555178 containerd[1569]: time="2026-03-06T02:34:46.554532051Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Mar 6 02:34:46.555178 containerd[1569]: time="2026-03-06T02:34:46.554621458Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.554682001Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.554828174Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.554856587Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.554980027Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.555000315Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.555017848Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.555031453Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.555044738Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.555174320Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.555195640Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Mar 6 02:34:46.557127 containerd[1569]: time="2026-03-06T02:34:46.555210728Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557586930Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557624090Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557638457Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557652504Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557663864Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557680625Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557704350Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557726421Z" level=info msg="runtime interface created" Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557734226Z" level=info msg="created NRI interface" Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557745026Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557762148Z" level=info msg="Connect containerd service" Mar 6 02:34:46.557978 containerd[1569]: time="2026-03-06T02:34:46.557793937Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 6 02:34:46.565585 containerd[1569]: time="2026-03-06T02:34:46.564479401Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 6 02:34:46.914706 tar[1562]: linux-amd64/README.md Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.978558774Z" level=info msg="Start subscribing containerd event" Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.979734860Z" level=info msg="Start recovering state" Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.979984857Z" level=info msg="Start event monitor" Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.980006948Z" level=info msg="Start cni network conf syncer for default" Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.980032245Z" level=info msg="Start streaming server" Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.980063714Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.980073984Z" level=info msg="runtime interface starting up..." Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.980082098Z" level=info msg="starting plugins..." Mar 6 02:34:46.980620 containerd[1569]: time="2026-03-06T02:34:46.980110702Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Mar 6 02:34:46.990565 containerd[1569]: time="2026-03-06T02:34:46.990243646Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 6 02:34:46.991262 containerd[1569]: time="2026-03-06T02:34:46.991239455Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 6 02:34:46.991782 systemd[1]: Started containerd.service - containerd container runtime. Mar 6 02:34:46.993736 containerd[1569]: time="2026-03-06T02:34:46.993708827Z" level=info msg="containerd successfully booted in 0.585471s" Mar 6 02:34:47.040806 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 6 02:34:47.317960 kernel: EDAC MC: Ver: 3.0.0 Mar 6 02:34:48.835569 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:34:48.861792 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 6 02:34:48.891553 (kubelet)[1680]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:34:48.897615 systemd[1]: Startup finished in 24.320s (kernel) + 42.001s (initrd) + 27.658s (userspace) = 1min 33.980s. Mar 6 02:34:49.483229 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 6 02:34:49.488614 systemd[1]: Started sshd@0-10.0.0.69:22-10.0.0.1:60938.service - OpenSSH per-connection server daemon (10.0.0.1:60938). Mar 6 02:34:49.817455 sshd[1692]: Accepted publickey for core from 10.0.0.1 port 60938 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:34:49.826580 sshd-session[1692]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:34:49.863058 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 6 02:34:49.866149 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 6 02:34:49.892224 systemd-logind[1544]: New session 1 of user core. Mar 6 02:34:49.944124 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 6 02:34:49.953509 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 6 02:34:49.991152 (systemd)[1697]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 6 02:34:50.016064 systemd-logind[1544]: New session c1 of user core. Mar 6 02:34:50.577018 systemd[1697]: Queued start job for default target default.target. Mar 6 02:34:50.601172 systemd[1697]: Created slice app.slice - User Application Slice. Mar 6 02:34:50.601624 systemd[1697]: Reached target paths.target - Paths. Mar 6 02:34:50.601807 systemd[1697]: Reached target timers.target - Timers. Mar 6 02:34:50.614494 systemd[1697]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 6 02:34:50.669968 systemd[1697]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 6 02:34:50.671094 systemd[1697]: Reached target sockets.target - Sockets. Mar 6 02:34:50.671625 systemd[1697]: Reached target basic.target - Basic System. Mar 6 02:34:50.671799 systemd[1697]: Reached target default.target - Main User Target. Mar 6 02:34:50.672263 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 6 02:34:50.674257 kubelet[1680]: E0306 02:34:50.673270 1680 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:34:50.672498 systemd[1697]: Startup finished in 627ms. Mar 6 02:34:50.698245 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 6 02:34:50.699144 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:34:50.700986 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:34:50.701668 systemd[1]: kubelet.service: Consumed 2.846s CPU time, 257M memory peak. Mar 6 02:34:50.766839 systemd[1]: Started sshd@1-10.0.0.69:22-10.0.0.1:60946.service - OpenSSH per-connection server daemon (10.0.0.1:60946). Mar 6 02:34:50.919252 sshd[1710]: Accepted publickey for core from 10.0.0.1 port 60946 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:34:50.926199 sshd-session[1710]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:34:50.968999 systemd-logind[1544]: New session 2 of user core. Mar 6 02:34:50.989251 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 6 02:34:51.063513 sshd[1713]: Connection closed by 10.0.0.1 port 60946 Mar 6 02:34:51.063155 sshd-session[1710]: pam_unix(sshd:session): session closed for user core Mar 6 02:34:51.085619 systemd[1]: sshd@1-10.0.0.69:22-10.0.0.1:60946.service: Deactivated successfully. Mar 6 02:34:51.093219 systemd[1]: session-2.scope: Deactivated successfully. Mar 6 02:34:51.102446 systemd-logind[1544]: Session 2 logged out. Waiting for processes to exit. Mar 6 02:34:51.108584 systemd[1]: Started sshd@2-10.0.0.69:22-10.0.0.1:60962.service - OpenSSH per-connection server daemon (10.0.0.1:60962). Mar 6 02:34:51.117618 systemd-logind[1544]: Removed session 2. Mar 6 02:34:51.270695 sshd[1719]: Accepted publickey for core from 10.0.0.1 port 60962 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:34:51.273211 sshd-session[1719]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:34:51.314164 systemd-logind[1544]: New session 3 of user core. Mar 6 02:34:51.330993 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 6 02:34:51.380517 sshd[1722]: Connection closed by 10.0.0.1 port 60962 Mar 6 02:34:51.379706 sshd-session[1719]: pam_unix(sshd:session): session closed for user core Mar 6 02:34:51.395808 systemd[1]: Started sshd@3-10.0.0.69:22-10.0.0.1:60966.service - OpenSSH per-connection server daemon (10.0.0.1:60966). Mar 6 02:34:51.398742 systemd[1]: sshd@2-10.0.0.69:22-10.0.0.1:60962.service: Deactivated successfully. Mar 6 02:34:51.405782 systemd[1]: session-3.scope: Deactivated successfully. Mar 6 02:34:51.411570 systemd-logind[1544]: Session 3 logged out. Waiting for processes to exit. Mar 6 02:34:51.428977 systemd-logind[1544]: Removed session 3. Mar 6 02:34:51.540242 sshd[1725]: Accepted publickey for core from 10.0.0.1 port 60966 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:34:51.548070 sshd-session[1725]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:34:51.586671 systemd-logind[1544]: New session 4 of user core. Mar 6 02:34:51.610707 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 6 02:34:51.683666 sshd[1731]: Connection closed by 10.0.0.1 port 60966 Mar 6 02:34:51.687975 sshd-session[1725]: pam_unix(sshd:session): session closed for user core Mar 6 02:34:51.708256 systemd[1]: sshd@3-10.0.0.69:22-10.0.0.1:60966.service: Deactivated successfully. Mar 6 02:34:51.714158 systemd[1]: session-4.scope: Deactivated successfully. Mar 6 02:34:51.716638 systemd-logind[1544]: Session 4 logged out. Waiting for processes to exit. Mar 6 02:34:51.725124 systemd[1]: Started sshd@4-10.0.0.69:22-10.0.0.1:60968.service - OpenSSH per-connection server daemon (10.0.0.1:60968). Mar 6 02:34:51.731180 systemd-logind[1544]: Removed session 4. Mar 6 02:34:51.866998 sshd[1737]: Accepted publickey for core from 10.0.0.1 port 60968 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:34:51.871978 sshd-session[1737]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:34:51.892743 systemd-logind[1544]: New session 5 of user core. Mar 6 02:34:51.909702 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 6 02:34:51.997036 sudo[1741]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 6 02:34:51.998149 sudo[1741]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 6 02:34:52.052807 sudo[1741]: pam_unix(sudo:session): session closed for user root Mar 6 02:34:52.062987 sshd[1740]: Connection closed by 10.0.0.1 port 60968 Mar 6 02:34:52.061762 sshd-session[1737]: pam_unix(sshd:session): session closed for user core Mar 6 02:34:52.082116 systemd[1]: sshd@4-10.0.0.69:22-10.0.0.1:60968.service: Deactivated successfully. Mar 6 02:34:52.085674 systemd[1]: session-5.scope: Deactivated successfully. Mar 6 02:34:52.088473 systemd-logind[1544]: Session 5 logged out. Waiting for processes to exit. Mar 6 02:34:52.093670 systemd[1]: Started sshd@5-10.0.0.69:22-10.0.0.1:38314.service - OpenSSH per-connection server daemon (10.0.0.1:38314). Mar 6 02:34:52.104214 systemd-logind[1544]: Removed session 5. Mar 6 02:34:52.222434 sshd[1747]: Accepted publickey for core from 10.0.0.1 port 38314 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:34:52.225747 sshd-session[1747]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:34:52.247626 systemd-logind[1544]: New session 6 of user core. Mar 6 02:34:52.260041 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 6 02:34:52.307052 sudo[1752]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 6 02:34:52.308060 sudo[1752]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 6 02:34:52.338271 sudo[1752]: pam_unix(sudo:session): session closed for user root Mar 6 02:34:52.365845 sudo[1751]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 6 02:34:52.366825 sudo[1751]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 6 02:34:52.407124 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 6 02:34:52.573106 augenrules[1774]: No rules Mar 6 02:34:52.577007 systemd[1]: audit-rules.service: Deactivated successfully. Mar 6 02:34:52.578036 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 6 02:34:52.581099 sudo[1751]: pam_unix(sudo:session): session closed for user root Mar 6 02:34:52.588691 sshd[1750]: Connection closed by 10.0.0.1 port 38314 Mar 6 02:34:52.589732 sshd-session[1747]: pam_unix(sshd:session): session closed for user core Mar 6 02:34:52.609449 systemd[1]: sshd@5-10.0.0.69:22-10.0.0.1:38314.service: Deactivated successfully. Mar 6 02:34:52.617221 systemd[1]: session-6.scope: Deactivated successfully. Mar 6 02:34:52.624200 systemd-logind[1544]: Session 6 logged out. Waiting for processes to exit. Mar 6 02:34:52.629494 systemd[1]: Started sshd@6-10.0.0.69:22-10.0.0.1:38322.service - OpenSSH per-connection server daemon (10.0.0.1:38322). Mar 6 02:34:52.637264 systemd-logind[1544]: Removed session 6. Mar 6 02:34:52.767603 sshd[1783]: Accepted publickey for core from 10.0.0.1 port 38322 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:34:52.774667 sshd-session[1783]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:34:52.808629 systemd-logind[1544]: New session 7 of user core. Mar 6 02:34:52.831202 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 6 02:34:52.897044 sudo[1787]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 6 02:34:52.898246 sudo[1787]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 6 02:34:55.510208 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 6 02:34:55.592164 (dockerd)[1809]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 6 02:35:01.116274 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 6 02:35:05.189760 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:35:07.780556 dockerd[1809]: time="2026-03-06T02:35:07.779648419Z" level=info msg="Starting up" Mar 6 02:35:07.795797 dockerd[1809]: time="2026-03-06T02:35:07.792593570Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Mar 6 02:35:08.426116 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:35:08.525697 (kubelet)[1838]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:35:08.569609 dockerd[1809]: time="2026-03-06T02:35:08.562240079Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Mar 6 02:35:09.372782 systemd[1]: var-lib-docker-metacopy\x2dcheck1403916877-merged.mount: Deactivated successfully. Mar 6 02:35:09.863653 dockerd[1809]: time="2026-03-06T02:35:09.861869511Z" level=info msg="Loading containers: start." Mar 6 02:35:10.201504 kernel: Initializing XFRM netlink socket Mar 6 02:35:10.436704 kubelet[1838]: E0306 02:35:10.435539 1838 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:35:10.465684 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:35:10.466051 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:35:10.471721 systemd[1]: kubelet.service: Consumed 3.629s CPU time, 110.8M memory peak. Mar 6 02:35:14.824983 systemd-networkd[1497]: docker0: Link UP Mar 6 02:35:14.889607 dockerd[1809]: time="2026-03-06T02:35:14.888968831Z" level=info msg="Loading containers: done." Mar 6 02:35:15.424490 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2357429941-merged.mount: Deactivated successfully. Mar 6 02:35:15.482040 dockerd[1809]: time="2026-03-06T02:35:15.479603530Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 6 02:35:15.488657 dockerd[1809]: time="2026-03-06T02:35:15.483646751Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Mar 6 02:35:15.498876 dockerd[1809]: time="2026-03-06T02:35:15.489491088Z" level=info msg="Initializing buildkit" Mar 6 02:35:16.138486 dockerd[1809]: time="2026-03-06T02:35:16.137059550Z" level=info msg="Completed buildkit initialization" Mar 6 02:35:16.302093 dockerd[1809]: time="2026-03-06T02:35:16.299929057Z" level=info msg="Daemon has completed initialization" Mar 6 02:35:16.302093 dockerd[1809]: time="2026-03-06T02:35:16.300817200Z" level=info msg="API listen on /run/docker.sock" Mar 6 02:35:16.301750 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 6 02:35:20.554975 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 6 02:35:20.594863 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:35:25.942980 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:35:26.501719 (kubelet)[2049]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:35:26.960661 update_engine[1551]: I20260306 02:35:26.946218 1551 update_attempter.cc:509] Updating boot flags... Mar 6 02:35:29.298100 kubelet[2049]: E0306 02:35:29.297024 2049 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:35:29.301650 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:35:29.301904 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:35:29.303192 systemd[1]: kubelet.service: Consumed 4.680s CPU time, 110.8M memory peak. Mar 6 02:35:29.764678 containerd[1569]: time="2026-03-06T02:35:29.755968176Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.35.2\"" Mar 6 02:35:33.325885 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1536180301.mount: Deactivated successfully. Mar 6 02:35:39.545135 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 6 02:35:39.556558 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:35:42.075951 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:35:42.100944 (kubelet)[2138]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:35:43.875916 kubelet[2138]: E0306 02:35:43.874844 2138 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:35:43.886705 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:35:43.887019 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:35:43.890901 systemd[1]: kubelet.service: Consumed 2.004s CPU time, 109M memory peak. Mar 6 02:35:54.047681 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 6 02:35:54.061619 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:35:56.973046 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:35:57.086549 (kubelet)[2158]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:35:58.542866 kubelet[2158]: E0306 02:35:58.538879 2158 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:35:58.561222 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:35:58.561927 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:35:58.564210 systemd[1]: kubelet.service: Consumed 2.078s CPU time, 111.1M memory peak. Mar 6 02:35:58.661739 containerd[1569]: time="2026-03-06T02:35:58.661002248Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.35.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:35:58.664089 containerd[1569]: time="2026-03-06T02:35:58.663109189Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.35.2: active requests=0, bytes read=27696467" Mar 6 02:35:58.669754 containerd[1569]: time="2026-03-06T02:35:58.669486373Z" level=info msg="ImageCreate event name:\"sha256:66108468ce51257077e642f2f509cd61d470029036a7954a1a47ca15b2706dda\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:35:58.677781 containerd[1569]: time="2026-03-06T02:35:58.677713561Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:68cdc586f13b13edb7aa30a18155be530136a39cfd5ef8672aad8ccc98f0a7f7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:35:58.680083 containerd[1569]: time="2026-03-06T02:35:58.679767327Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.35.2\" with image id \"sha256:66108468ce51257077e642f2f509cd61d470029036a7954a1a47ca15b2706dda\", repo tag \"registry.k8s.io/kube-apiserver:v1.35.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:68cdc586f13b13edb7aa30a18155be530136a39cfd5ef8672aad8ccc98f0a7f7\", size \"27693066\" in 28.9226267s" Mar 6 02:35:58.680083 containerd[1569]: time="2026-03-06T02:35:58.680048222Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.35.2\" returns image reference \"sha256:66108468ce51257077e642f2f509cd61d470029036a7954a1a47ca15b2706dda\"" Mar 6 02:35:58.691268 containerd[1569]: time="2026-03-06T02:35:58.689828136Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.35.2\"" Mar 6 02:36:08.302104 containerd[1569]: time="2026-03-06T02:36:08.300696314Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.35.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:08.306050 containerd[1569]: time="2026-03-06T02:36:08.304528771Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.35.2: active requests=0, bytes read=21450700" Mar 6 02:36:08.308054 containerd[1569]: time="2026-03-06T02:36:08.307844881Z" level=info msg="ImageCreate event name:\"sha256:0f2dd35011c05b55a97c9304ae1d36cfd58499cc1fd3dd8ccdf6efef1144e36a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:08.314703 containerd[1569]: time="2026-03-06T02:36:08.314197599Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:d9784320a41dd1b155c0ad8fdb5823d60c475870f3dd23865edde36b585748f2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:08.316070 containerd[1569]: time="2026-03-06T02:36:08.315856159Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.35.2\" with image id \"sha256:0f2dd35011c05b55a97c9304ae1d36cfd58499cc1fd3dd8ccdf6efef1144e36a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.35.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:d9784320a41dd1b155c0ad8fdb5823d60c475870f3dd23865edde36b585748f2\", size \"23142311\" in 9.62598884s" Mar 6 02:36:08.316198 containerd[1569]: time="2026-03-06T02:36:08.316144081Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.35.2\" returns image reference \"sha256:0f2dd35011c05b55a97c9304ae1d36cfd58499cc1fd3dd8ccdf6efef1144e36a\"" Mar 6 02:36:08.323195 containerd[1569]: time="2026-03-06T02:36:08.323034243Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.35.2\"" Mar 6 02:36:08.794841 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Mar 6 02:36:08.807919 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:36:09.348113 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:36:09.373942 (kubelet)[2178]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:36:09.523817 kubelet[2178]: E0306 02:36:09.523536 2178 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:36:09.529970 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:36:09.530887 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:36:09.531816 systemd[1]: kubelet.service: Consumed 533ms CPU time, 108.8M memory peak. Mar 6 02:36:14.076222 containerd[1569]: time="2026-03-06T02:36:14.075694463Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.35.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:14.079558 containerd[1569]: time="2026-03-06T02:36:14.079458349Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.35.2: active requests=0, bytes read=15548429" Mar 6 02:36:14.083434 containerd[1569]: time="2026-03-06T02:36:14.082980666Z" level=info msg="ImageCreate event name:\"sha256:ee83c410d7938aa1752b4e79a8d51f03710b4becc23b2e095fba471049fb2914\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:14.094005 containerd[1569]: time="2026-03-06T02:36:14.093797131Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:5833e2c4b779215efe7a48126c067de199e86aa5a86518693adeef16db0ff943\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:14.099003 containerd[1569]: time="2026-03-06T02:36:14.098014854Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.35.2\" with image id \"sha256:ee83c410d7938aa1752b4e79a8d51f03710b4becc23b2e095fba471049fb2914\", repo tag \"registry.k8s.io/kube-scheduler:v1.35.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:5833e2c4b779215efe7a48126c067de199e86aa5a86518693adeef16db0ff943\", size \"17240058\" in 5.774864686s" Mar 6 02:36:14.099003 containerd[1569]: time="2026-03-06T02:36:14.098167867Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.35.2\" returns image reference \"sha256:ee83c410d7938aa1752b4e79a8d51f03710b4becc23b2e095fba471049fb2914\"" Mar 6 02:36:14.120103 containerd[1569]: time="2026-03-06T02:36:14.119882342Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.35.2\"" Mar 6 02:36:19.586976 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Mar 6 02:36:19.654816 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:36:20.270559 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1032773255.mount: Deactivated successfully. Mar 6 02:36:20.384216 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:36:20.432134 (kubelet)[2204]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:36:23.781094 kubelet[2204]: E0306 02:36:23.780497 2204 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:36:23.795884 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:36:23.796243 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:36:23.797073 systemd[1]: kubelet.service: Consumed 2.993s CPU time, 110.6M memory peak. Mar 6 02:36:26.675582 containerd[1569]: time="2026-03-06T02:36:26.674528314Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.35.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:26.682123 containerd[1569]: time="2026-03-06T02:36:26.679015413Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.35.2: active requests=0, bytes read=25685312" Mar 6 02:36:26.688567 containerd[1569]: time="2026-03-06T02:36:26.688237644Z" level=info msg="ImageCreate event name:\"sha256:3c471cf273e44f68c91b48985c27627d581915b9ee5e72f7227bbf2146008b5e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:26.716172 containerd[1569]: time="2026-03-06T02:36:26.709991356Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:015265214cc874b593a7adccdcfe4ac15d2b8e9ae89881bdcd5bcb99d42e1862\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:26.719120 containerd[1569]: time="2026-03-06T02:36:26.718600218Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.35.2\" with image id \"sha256:3c471cf273e44f68c91b48985c27627d581915b9ee5e72f7227bbf2146008b5e\", repo tag \"registry.k8s.io/kube-proxy:v1.35.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:015265214cc874b593a7adccdcfe4ac15d2b8e9ae89881bdcd5bcb99d42e1862\", size \"25684331\" in 12.598581052s" Mar 6 02:36:26.719120 containerd[1569]: time="2026-03-06T02:36:26.718642837Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.35.2\" returns image reference \"sha256:3c471cf273e44f68c91b48985c27627d581915b9ee5e72f7227bbf2146008b5e\"" Mar 6 02:36:26.730498 containerd[1569]: time="2026-03-06T02:36:26.729055765Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.13.1\"" Mar 6 02:36:28.166595 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3620492696.mount: Deactivated successfully. Mar 6 02:36:34.169103 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Mar 6 02:36:34.211106 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:36:35.835109 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:36:35.915778 (kubelet)[2280]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:36:36.750950 kubelet[2280]: E0306 02:36:36.750024 2280 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:36:36.760187 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:36:36.760907 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:36:36.763592 systemd[1]: kubelet.service: Consumed 1.300s CPU time, 110.5M memory peak. Mar 6 02:36:40.609869 containerd[1569]: time="2026-03-06T02:36:40.608094102Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.13.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:40.614263 containerd[1569]: time="2026-03-06T02:36:40.612667770Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.13.1: active requests=0, bytes read=23556542" Mar 6 02:36:40.620070 containerd[1569]: time="2026-03-06T02:36:40.620026710Z" level=info msg="ImageCreate event name:\"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:40.644213 containerd[1569]: time="2026-03-06T02:36:40.644093918Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9b9128672209474da07c91439bf15ed704ae05ad918dd6454e5b6ae14e35fee6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:40.653921 containerd[1569]: time="2026-03-06T02:36:40.653268138Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.13.1\" with image id \"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\", repo tag \"registry.k8s.io/coredns/coredns:v1.13.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9b9128672209474da07c91439bf15ed704ae05ad918dd6454e5b6ae14e35fee6\", size \"23553139\" in 13.924165667s" Mar 6 02:36:40.654889 containerd[1569]: time="2026-03-06T02:36:40.654222933Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.13.1\" returns image reference \"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\"" Mar 6 02:36:40.668714 containerd[1569]: time="2026-03-06T02:36:40.665153516Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Mar 6 02:36:42.230576 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3484076019.mount: Deactivated successfully. Mar 6 02:36:42.272792 containerd[1569]: time="2026-03-06T02:36:42.272252726Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:42.279992 containerd[1569]: time="2026-03-06T02:36:42.278858981Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=321218" Mar 6 02:36:42.286934 containerd[1569]: time="2026-03-06T02:36:42.285588186Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:42.292640 containerd[1569]: time="2026-03-06T02:36:42.291715757Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:42.293622 containerd[1569]: time="2026-03-06T02:36:42.293249750Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 1.628040429s" Mar 6 02:36:42.296610 containerd[1569]: time="2026-03-06T02:36:42.293587521Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Mar 6 02:36:42.317531 containerd[1569]: time="2026-03-06T02:36:42.316637298Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.6-0\"" Mar 6 02:36:43.579985 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2213727130.mount: Deactivated successfully. Mar 6 02:36:46.798757 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Mar 6 02:36:46.973660 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:36:47.740513 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:36:47.761852 (kubelet)[2349]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 02:36:48.145596 kubelet[2349]: E0306 02:36:48.144253 2349 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 02:36:48.153167 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 02:36:48.154149 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 02:36:48.155091 systemd[1]: kubelet.service: Consumed 640ms CPU time, 110.7M memory peak. Mar 6 02:36:52.269826 containerd[1569]: time="2026-03-06T02:36:52.263736619Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.6-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:52.285756 containerd[1569]: time="2026-03-06T02:36:52.273549718Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.6-0: active requests=0, bytes read=23630322" Mar 6 02:36:52.303146 containerd[1569]: time="2026-03-06T02:36:52.297208728Z" level=info msg="ImageCreate event name:\"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:52.402193 containerd[1569]: time="2026-03-06T02:36:52.396267613Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:60a30b5d81b2217555e2cfb9537f655b7ba97220b99c39ee2e162a7127225890\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:36:52.431128 containerd[1569]: time="2026-03-06T02:36:52.409142027Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.6-0\" with image id \"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\", repo tag \"registry.k8s.io/etcd:3.6.6-0\", repo digest \"registry.k8s.io/etcd@sha256:60a30b5d81b2217555e2cfb9537f655b7ba97220b99c39ee2e162a7127225890\", size \"23641797\" in 10.092376589s" Mar 6 02:36:52.431128 containerd[1569]: time="2026-03-06T02:36:52.421046620Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.6-0\" returns image reference \"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\"" Mar 6 02:36:57.587187 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:36:57.587598 systemd[1]: kubelet.service: Consumed 640ms CPU time, 110.7M memory peak. Mar 6 02:36:57.598652 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:36:57.800249 systemd[1]: Reload requested from client PID 2405 ('systemctl') (unit session-7.scope)... Mar 6 02:36:57.800271 systemd[1]: Reloading... Mar 6 02:36:58.089745 zram_generator::config[2445]: No configuration found. Mar 6 02:36:58.767977 systemd[1]: Reloading finished in 966 ms. Mar 6 02:36:59.030804 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:36:59.056738 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:36:59.074232 systemd[1]: kubelet.service: Deactivated successfully. Mar 6 02:36:59.075242 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:36:59.075553 systemd[1]: kubelet.service: Consumed 292ms CPU time, 98.4M memory peak. Mar 6 02:36:59.086053 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:37:00.338638 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:37:00.406270 (kubelet)[2497]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 6 02:37:01.089803 kubelet[2497]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 6 02:37:01.694762 kubelet[2497]: I0306 02:37:01.685236 2497 server.go:525] "Kubelet version" kubeletVersion="v1.35.1" Mar 6 02:37:01.694762 kubelet[2497]: I0306 02:37:01.694659 2497 server.go:527] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 6 02:37:01.700081 kubelet[2497]: I0306 02:37:01.695190 2497 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 6 02:37:01.700081 kubelet[2497]: I0306 02:37:01.695209 2497 watchdog_linux.go:138] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 6 02:37:01.700081 kubelet[2497]: I0306 02:37:01.697261 2497 server.go:951] "Client rotation is on, will bootstrap in background" Mar 6 02:37:01.869901 kubelet[2497]: E0306 02:37:01.868702 2497 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.69:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 6 02:37:01.869901 kubelet[2497]: I0306 02:37:01.869767 2497 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 6 02:37:02.026054 kubelet[2497]: I0306 02:37:02.023958 2497 server.go:1418] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 6 02:37:02.065564 kubelet[2497]: I0306 02:37:02.063969 2497 server.go:775] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 6 02:37:02.069291 kubelet[2497]: I0306 02:37:02.069113 2497 container_manager_linux.go:272] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 6 02:37:02.071767 kubelet[2497]: I0306 02:37:02.069268 2497 container_manager_linux.go:277] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 6 02:37:02.075609 kubelet[2497]: I0306 02:37:02.071914 2497 topology_manager.go:143] "Creating topology manager with none policy" Mar 6 02:37:02.075609 kubelet[2497]: I0306 02:37:02.071936 2497 container_manager_linux.go:308] "Creating device plugin manager" Mar 6 02:37:02.075609 kubelet[2497]: I0306 02:37:02.072269 2497 container_manager_linux.go:317] "Creating Dynamic Resource Allocation (DRA) manager" Mar 6 02:37:02.081791 kubelet[2497]: I0306 02:37:02.081238 2497 state_mem.go:41] "Initialized" logger="CPUManager state memory" Mar 6 02:37:02.082738 kubelet[2497]: I0306 02:37:02.082227 2497 kubelet.go:482] "Attempting to sync node with API server" Mar 6 02:37:02.084576 kubelet[2497]: I0306 02:37:02.083829 2497 kubelet.go:383] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 6 02:37:02.084576 kubelet[2497]: I0306 02:37:02.083963 2497 kubelet.go:394] "Adding apiserver pod source" Mar 6 02:37:02.084576 kubelet[2497]: I0306 02:37:02.084069 2497 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 6 02:37:02.146718 kubelet[2497]: I0306 02:37:02.145820 2497 kuberuntime_manager.go:294] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Mar 6 02:37:02.154226 kubelet[2497]: I0306 02:37:02.154066 2497 kubelet.go:943] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 6 02:37:02.154226 kubelet[2497]: I0306 02:37:02.154202 2497 kubelet.go:970] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 6 02:37:02.155769 kubelet[2497]: W0306 02:37:02.154978 2497 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 6 02:37:02.188753 kubelet[2497]: I0306 02:37:02.187734 2497 server.go:1257] "Started kubelet" Mar 6 02:37:02.197130 kubelet[2497]: I0306 02:37:02.190640 2497 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 6 02:37:02.197130 kubelet[2497]: I0306 02:37:02.195875 2497 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 6 02:37:02.197130 kubelet[2497]: I0306 02:37:02.196177 2497 server.go:254] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 6 02:37:02.197130 kubelet[2497]: I0306 02:37:02.196198 2497 server.go:182] "Starting to listen" address="0.0.0.0" port=10250 Mar 6 02:37:02.203957 kubelet[2497]: I0306 02:37:02.200207 2497 fs_resource_analyzer.go:69] "Starting FS ResourceAnalyzer" Mar 6 02:37:02.203957 kubelet[2497]: I0306 02:37:02.203829 2497 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 6 02:37:02.228655 kubelet[2497]: I0306 02:37:02.226872 2497 server.go:317] "Adding debug handlers to kubelet server" Mar 6 02:37:02.253729 kubelet[2497]: I0306 02:37:02.253185 2497 volume_manager.go:311] "Starting Kubelet Volume Manager" Mar 6 02:37:02.253729 kubelet[2497]: E0306 02:37:02.253721 2497 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 02:37:02.255071 kubelet[2497]: I0306 02:37:02.254694 2497 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 6 02:37:02.255231 kubelet[2497]: I0306 02:37:02.255131 2497 reconciler.go:29] "Reconciler: start to sync state" Mar 6 02:37:02.256889 kubelet[2497]: E0306 02:37:02.256755 2497 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="200ms" Mar 6 02:37:02.259189 kubelet[2497]: E0306 02:37:02.259072 2497 kubelet.go:1656] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 6 02:37:02.266711 kubelet[2497]: E0306 02:37:02.260183 2497 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.69:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.69:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.189a200807d84cea default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-06 02:37:02.187154666 +0000 UTC m=+1.696321689,LastTimestamp:2026-03-06 02:37:02.187154666 +0000 UTC m=+1.696321689,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 6 02:37:02.272686 kubelet[2497]: I0306 02:37:02.272267 2497 factory.go:223] Registration of the systemd container factory successfully Mar 6 02:37:02.275912 kubelet[2497]: I0306 02:37:02.275711 2497 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 6 02:37:02.289787 kubelet[2497]: I0306 02:37:02.289144 2497 factory.go:223] Registration of the containerd container factory successfully Mar 6 02:37:02.304187 kubelet[2497]: I0306 02:37:02.303123 2497 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 6 02:37:02.354243 kubelet[2497]: E0306 02:37:02.354160 2497 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 02:37:02.385099 kubelet[2497]: I0306 02:37:02.384952 2497 cpu_manager.go:225] "Starting" policy="none" Mar 6 02:37:02.385099 kubelet[2497]: I0306 02:37:02.385075 2497 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 6 02:37:02.385099 kubelet[2497]: I0306 02:37:02.385107 2497 state_mem.go:41] "Initialized" logger="CPUManager state checkpoint.CPUManager state memory" Mar 6 02:37:02.396942 kubelet[2497]: I0306 02:37:02.396207 2497 policy_none.go:50] "Start" Mar 6 02:37:02.397231 kubelet[2497]: I0306 02:37:02.397102 2497 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 6 02:37:02.397667 kubelet[2497]: I0306 02:37:02.397236 2497 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 6 02:37:02.422822 kubelet[2497]: I0306 02:37:02.421024 2497 policy_none.go:44] "Start" Mar 6 02:37:02.454819 kubelet[2497]: E0306 02:37:02.454670 2497 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 02:37:02.454819 kubelet[2497]: I0306 02:37:02.454801 2497 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 6 02:37:02.456674 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 6 02:37:02.460041 kubelet[2497]: I0306 02:37:02.456915 2497 status_manager.go:249] "Starting to sync pod status with apiserver" Mar 6 02:37:02.461989 kubelet[2497]: I0306 02:37:02.461794 2497 kubelet.go:2501] "Starting kubelet main sync loop" Mar 6 02:37:02.462066 kubelet[2497]: E0306 02:37:02.462017 2497 kubelet.go:2525] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 6 02:37:02.464962 kubelet[2497]: E0306 02:37:02.464843 2497 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="400ms" Mar 6 02:37:02.502918 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 6 02:37:02.564926 kubelet[2497]: E0306 02:37:02.563112 2497 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 02:37:02.566679 kubelet[2497]: E0306 02:37:02.563759 2497 kubelet.go:2525] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 6 02:37:02.643253 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 6 02:37:02.682070 kubelet[2497]: E0306 02:37:02.674100 2497 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 02:37:02.701224 kubelet[2497]: E0306 02:37:02.699544 2497 manager.go:525] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 6 02:37:02.701224 kubelet[2497]: I0306 02:37:02.700887 2497 eviction_manager.go:194] "Eviction manager: starting control loop" Mar 6 02:37:02.701224 kubelet[2497]: I0306 02:37:02.700909 2497 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 6 02:37:02.727053 kubelet[2497]: I0306 02:37:02.722125 2497 plugin_manager.go:121] "Starting Kubelet Plugin Manager" Mar 6 02:37:02.758687 kubelet[2497]: E0306 02:37:02.758141 2497 eviction_manager.go:272] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 6 02:37:02.760737 kubelet[2497]: E0306 02:37:02.760221 2497 eviction_manager.go:297] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 6 02:37:02.876112 kubelet[2497]: E0306 02:37:02.867108 2497 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="800ms" Mar 6 02:37:02.881721 kubelet[2497]: I0306 02:37:02.881114 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7e99612e9106acc372b437b601cabe3e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"7e99612e9106acc372b437b601cabe3e\") " pod="kube-system/kube-apiserver-localhost" Mar 6 02:37:02.881721 kubelet[2497]: I0306 02:37:02.881164 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7e99612e9106acc372b437b601cabe3e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"7e99612e9106acc372b437b601cabe3e\") " pod="kube-system/kube-apiserver-localhost" Mar 6 02:37:02.881721 kubelet[2497]: I0306 02:37:02.881192 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7e99612e9106acc372b437b601cabe3e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"7e99612e9106acc372b437b601cabe3e\") " pod="kube-system/kube-apiserver-localhost" Mar 6 02:37:02.919131 kubelet[2497]: I0306 02:37:02.916692 2497 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Mar 6 02:37:02.922274 kubelet[2497]: E0306 02:37:02.921767 2497 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Mar 6 02:37:03.041990 kubelet[2497]: I0306 02:37:03.041871 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:03.047889 kubelet[2497]: I0306 02:37:03.047159 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:03.047889 kubelet[2497]: I0306 02:37:03.047216 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:03.047889 kubelet[2497]: I0306 02:37:03.047243 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:03.047889 kubelet[2497]: I0306 02:37:03.047598 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:03.047889 kubelet[2497]: I0306 02:37:03.047628 2497 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bd81bb6a14e176da833e3a8030ee5eac-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"bd81bb6a14e176da833e3a8030ee5eac\") " pod="kube-system/kube-scheduler-localhost" Mar 6 02:37:03.059216 systemd[1]: Created slice kubepods-burstable-pod7e99612e9106acc372b437b601cabe3e.slice - libcontainer container kubepods-burstable-pod7e99612e9106acc372b437b601cabe3e.slice. Mar 6 02:37:03.154088 kubelet[2497]: I0306 02:37:03.148048 2497 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Mar 6 02:37:03.160075 kubelet[2497]: E0306 02:37:03.156888 2497 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Mar 6 02:37:03.160075 kubelet[2497]: E0306 02:37:03.159234 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:03.578962 containerd[1569]: time="2026-03-06T02:37:03.392271092Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:7e99612e9106acc372b437b601cabe3e,Namespace:kube-system,Attempt:0,}" Mar 6 02:37:03.582483 kubelet[2497]: I0306 02:37:03.581264 2497 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Mar 6 02:37:03.582483 kubelet[2497]: E0306 02:37:03.582136 2497 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Mar 6 02:37:04.378010 systemd[1]: Created slice kubepods-burstable-podf420dd303687d038b2bc2fa1d277c55c.slice - libcontainer container kubepods-burstable-podf420dd303687d038b2bc2fa1d277c55c.slice. Mar 6 02:37:04.393050 kubelet[2497]: E0306 02:37:04.388690 2497 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="1.6s" Mar 6 02:37:04.469065 kubelet[2497]: E0306 02:37:04.396048 2497 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.69:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 6 02:37:04.554753 kubelet[2497]: E0306 02:37:04.545246 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:04.744013 kubelet[2497]: I0306 02:37:04.692145 2497 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Mar 6 02:37:04.744013 kubelet[2497]: E0306 02:37:04.743009 2497 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Mar 6 02:37:04.752130 containerd[1569]: time="2026-03-06T02:37:04.745133837Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:f420dd303687d038b2bc2fa1d277c55c,Namespace:kube-system,Attempt:0,}" Mar 6 02:37:05.286854 systemd[1]: Created slice kubepods-burstable-podbd81bb6a14e176da833e3a8030ee5eac.slice - libcontainer container kubepods-burstable-podbd81bb6a14e176da833e3a8030ee5eac.slice. Mar 6 02:37:05.676050 kubelet[2497]: E0306 02:37:05.675781 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:06.165782 kubelet[2497]: E0306 02:37:06.161224 2497 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="3.2s" Mar 6 02:37:06.696679 containerd[1569]: time="2026-03-06T02:37:06.692818718Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:bd81bb6a14e176da833e3a8030ee5eac,Namespace:kube-system,Attempt:0,}" Mar 6 02:37:07.941113 kubelet[2497]: I0306 02:37:07.940090 2497 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Mar 6 02:37:07.941113 kubelet[2497]: E0306 02:37:07.941022 2497 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Mar 6 02:37:08.435225 kubelet[2497]: E0306 02:37:08.435153 2497 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.69:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 6 02:37:08.946845 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3751326356.mount: Deactivated successfully. Mar 6 02:37:08.989633 containerd[1569]: time="2026-03-06T02:37:08.988044253Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 02:37:09.012994 containerd[1569]: time="2026-03-06T02:37:09.012944847Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Mar 6 02:37:09.023253 containerd[1569]: time="2026-03-06T02:37:09.022808552Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 02:37:09.028978 containerd[1569]: time="2026-03-06T02:37:09.027818478Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 02:37:09.031201 containerd[1569]: time="2026-03-06T02:37:09.030714940Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Mar 6 02:37:09.045900 containerd[1569]: time="2026-03-06T02:37:09.044844242Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 02:37:09.049070 containerd[1569]: time="2026-03-06T02:37:09.048238261Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Mar 6 02:37:09.056949 containerd[1569]: time="2026-03-06T02:37:09.056725837Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 02:37:09.058765 containerd[1569]: time="2026-03-06T02:37:09.058159636Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 5.012110751s" Mar 6 02:37:09.064153 containerd[1569]: time="2026-03-06T02:37:09.063082141Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 3.91018747s" Mar 6 02:37:09.083744 containerd[1569]: time="2026-03-06T02:37:09.081677877Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 1.147579412s" Mar 6 02:37:09.376726 kubelet[2497]: E0306 02:37:09.375745 2497 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="6.4s" Mar 6 02:37:09.485724 containerd[1569]: time="2026-03-06T02:37:09.484745271Z" level=info msg="connecting to shim 741c5061cf0493872b8c57c8cfa830ad32da9b3f36c45f395b83146dce45632e" address="unix:///run/containerd/s/d18e13d50c3488003a3699537291b7da069c0b3de7dae18de4bd7882f45e92c0" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:37:09.497915 containerd[1569]: time="2026-03-06T02:37:09.497731713Z" level=info msg="connecting to shim a85e44ca8b894f7dcb7b4bf2b032a3e44a8bf6de4907c99ea6a9974e4e12bba9" address="unix:///run/containerd/s/6eba47555f86d3da43939a22980904f6362cad0adba4c1bb97fd4f1b62c1c242" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:37:09.556215 containerd[1569]: time="2026-03-06T02:37:09.555968726Z" level=info msg="connecting to shim 76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299" address="unix:///run/containerd/s/34a31aeda2ae590da7b89310a1b259a615a310a1e045c82ab392d45f799ab99c" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:37:10.185129 systemd[1]: Started cri-containerd-741c5061cf0493872b8c57c8cfa830ad32da9b3f36c45f395b83146dce45632e.scope - libcontainer container 741c5061cf0493872b8c57c8cfa830ad32da9b3f36c45f395b83146dce45632e. Mar 6 02:37:10.269749 systemd[1]: Started cri-containerd-a85e44ca8b894f7dcb7b4bf2b032a3e44a8bf6de4907c99ea6a9974e4e12bba9.scope - libcontainer container a85e44ca8b894f7dcb7b4bf2b032a3e44a8bf6de4907c99ea6a9974e4e12bba9. Mar 6 02:37:10.423693 systemd[1]: Started cri-containerd-76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299.scope - libcontainer container 76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299. Mar 6 02:37:10.690735 kubelet[2497]: E0306 02:37:10.688754 2497 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.69:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.69:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.189a200807d84cea default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-06 02:37:02.187154666 +0000 UTC m=+1.696321689,LastTimestamp:2026-03-06 02:37:02.187154666 +0000 UTC m=+1.696321689,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 6 02:37:11.093960 containerd[1569]: time="2026-03-06T02:37:11.093907904Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:f420dd303687d038b2bc2fa1d277c55c,Namespace:kube-system,Attempt:0,} returns sandbox id \"a85e44ca8b894f7dcb7b4bf2b032a3e44a8bf6de4907c99ea6a9974e4e12bba9\"" Mar 6 02:37:11.149261 containerd[1569]: time="2026-03-06T02:37:11.149209326Z" level=info msg="CreateContainer within sandbox \"a85e44ca8b894f7dcb7b4bf2b032a3e44a8bf6de4907c99ea6a9974e4e12bba9\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 6 02:37:11.154206 kubelet[2497]: I0306 02:37:11.153798 2497 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Mar 6 02:37:11.155825 kubelet[2497]: E0306 02:37:11.155180 2497 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Mar 6 02:37:11.186207 containerd[1569]: time="2026-03-06T02:37:11.186162452Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:7e99612e9106acc372b437b601cabe3e,Namespace:kube-system,Attempt:0,} returns sandbox id \"741c5061cf0493872b8c57c8cfa830ad32da9b3f36c45f395b83146dce45632e\"" Mar 6 02:37:11.231748 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1236976084.mount: Deactivated successfully. Mar 6 02:37:11.246750 containerd[1569]: time="2026-03-06T02:37:11.245792471Z" level=info msg="CreateContainer within sandbox \"741c5061cf0493872b8c57c8cfa830ad32da9b3f36c45f395b83146dce45632e\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 6 02:37:11.246750 containerd[1569]: time="2026-03-06T02:37:11.246039507Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:bd81bb6a14e176da833e3a8030ee5eac,Namespace:kube-system,Attempt:0,} returns sandbox id \"76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299\"" Mar 6 02:37:11.257665 containerd[1569]: time="2026-03-06T02:37:11.256782917Z" level=info msg="Container bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:37:11.260092 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1618555173.mount: Deactivated successfully. Mar 6 02:37:11.290061 containerd[1569]: time="2026-03-06T02:37:11.287939836Z" level=info msg="CreateContainer within sandbox \"76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 6 02:37:11.378795 containerd[1569]: time="2026-03-06T02:37:11.375788830Z" level=info msg="CreateContainer within sandbox \"a85e44ca8b894f7dcb7b4bf2b032a3e44a8bf6de4907c99ea6a9974e4e12bba9\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd\"" Mar 6 02:37:11.378795 containerd[1569]: time="2026-03-06T02:37:11.376833340Z" level=info msg="Container 2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:37:11.380166 containerd[1569]: time="2026-03-06T02:37:11.379945153Z" level=info msg="StartContainer for \"bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd\"" Mar 6 02:37:11.398788 containerd[1569]: time="2026-03-06T02:37:11.398202196Z" level=info msg="connecting to shim bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd" address="unix:///run/containerd/s/6eba47555f86d3da43939a22980904f6362cad0adba4c1bb97fd4f1b62c1c242" protocol=ttrpc version=3 Mar 6 02:37:11.423891 containerd[1569]: time="2026-03-06T02:37:11.423135601Z" level=info msg="CreateContainer within sandbox \"741c5061cf0493872b8c57c8cfa830ad32da9b3f36c45f395b83146dce45632e\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695\"" Mar 6 02:37:11.428778 containerd[1569]: time="2026-03-06T02:37:11.428157655Z" level=info msg="StartContainer for \"2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695\"" Mar 6 02:37:11.438644 containerd[1569]: time="2026-03-06T02:37:11.438020519Z" level=info msg="connecting to shim 2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695" address="unix:///run/containerd/s/d18e13d50c3488003a3699537291b7da069c0b3de7dae18de4bd7882f45e92c0" protocol=ttrpc version=3 Mar 6 02:37:11.444028 containerd[1569]: time="2026-03-06T02:37:11.443747427Z" level=info msg="Container abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:37:11.501914 containerd[1569]: time="2026-03-06T02:37:11.501050907Z" level=info msg="CreateContainer within sandbox \"76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075\"" Mar 6 02:37:11.506901 containerd[1569]: time="2026-03-06T02:37:11.506655921Z" level=info msg="StartContainer for \"abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075\"" Mar 6 02:37:11.507200 systemd[1]: Started cri-containerd-bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd.scope - libcontainer container bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd. Mar 6 02:37:11.526721 containerd[1569]: time="2026-03-06T02:37:11.525741022Z" level=info msg="connecting to shim abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075" address="unix:///run/containerd/s/34a31aeda2ae590da7b89310a1b259a615a310a1e045c82ab392d45f799ab99c" protocol=ttrpc version=3 Mar 6 02:37:11.532650 systemd[1]: Started cri-containerd-2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695.scope - libcontainer container 2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695. Mar 6 02:37:11.738913 systemd[1]: Started cri-containerd-abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075.scope - libcontainer container abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075. Mar 6 02:37:11.858036 containerd[1569]: time="2026-03-06T02:37:11.857987355Z" level=info msg="StartContainer for \"bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd\" returns successfully" Mar 6 02:37:12.027937 containerd[1569]: time="2026-03-06T02:37:12.027027692Z" level=info msg="StartContainer for \"2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695\" returns successfully" Mar 6 02:37:12.034729 containerd[1569]: time="2026-03-06T02:37:12.034161240Z" level=info msg="StartContainer for \"abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075\" returns successfully" Mar 6 02:37:12.092885 kubelet[2497]: E0306 02:37:12.090849 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:12.140727 kubelet[2497]: E0306 02:37:12.139740 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:12.151257 kubelet[2497]: E0306 02:37:12.151233 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:12.215037 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3947106812.mount: Deactivated successfully. Mar 6 02:37:12.761997 kubelet[2497]: E0306 02:37:12.761754 2497 eviction_manager.go:297] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 6 02:37:13.157199 kubelet[2497]: E0306 02:37:13.156896 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:13.157199 kubelet[2497]: E0306 02:37:13.156956 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:14.183206 kubelet[2497]: E0306 02:37:14.182881 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:15.299522 kubelet[2497]: E0306 02:37:15.299050 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:16.488532 kubelet[2497]: E0306 02:37:16.487003 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:17.572619 kubelet[2497]: I0306 02:37:17.569757 2497 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Mar 6 02:37:19.838057 kubelet[2497]: E0306 02:37:19.836997 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:22.755039 kubelet[2497]: E0306 02:37:22.751653 2497 nodelease.go:50] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Mar 6 02:37:22.771212 kubelet[2497]: E0306 02:37:22.762639 2497 eviction_manager.go:297] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 6 02:37:27.354909 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 4598545924 wd_nsec: 4598544249 Mar 6 02:37:27.591548 kubelet[2497]: E0306 02:37:27.589721 2497 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 02:37:28.045844 kubelet[2497]: I0306 02:37:28.044631 2497 kubelet_node_status.go:77] "Successfully registered node" node="localhost" Mar 6 02:37:28.065959 kubelet[2497]: E0306 02:37:28.052755 2497 csi_plugin.go:399] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "localhost" not found Mar 6 02:37:28.065959 kubelet[2497]: I0306 02:37:28.054097 2497 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 6 02:37:28.135057 kubelet[2497]: E0306 02:37:28.134216 2497 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.189a200807d84cea default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-06 02:37:02.187154666 +0000 UTC m=+1.696321689,LastTimestamp:2026-03-06 02:37:02.187154666 +0000 UTC m=+1.696321689,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 6 02:37:28.339097 kubelet[2497]: E0306 02:37:28.337670 2497 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.189a20080c214eeb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-06 02:37:02.259048171 +0000 UTC m=+1.768215203,LastTimestamp:2026-03-06 02:37:02.259048171 +0000 UTC m=+1.768215203,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 6 02:37:28.378975 kubelet[2497]: I0306 02:37:28.376747 2497 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:28.393934 kubelet[2497]: I0306 02:37:28.392844 2497 apiserver.go:52] "Watching apiserver" Mar 6 02:37:28.500885 kubelet[2497]: I0306 02:37:28.499048 2497 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 6 02:37:28.658057 kubelet[2497]: I0306 02:37:28.656253 2497 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 6 02:37:29.877149 kubelet[2497]: I0306 02:37:29.873183 2497 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.873164906 podStartE2EDuration="1.873164906s" podCreationTimestamp="2026-03-06 02:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 02:37:29.870099319 +0000 UTC m=+29.379266362" watchObservedRunningTime="2026-03-06 02:37:29.873164906 +0000 UTC m=+29.382331938" Mar 6 02:37:29.919148 kubelet[2497]: I0306 02:37:29.919068 2497 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.919047427 podStartE2EDuration="1.919047427s" podCreationTimestamp="2026-03-06 02:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 02:37:29.915728689 +0000 UTC m=+29.424895731" watchObservedRunningTime="2026-03-06 02:37:29.919047427 +0000 UTC m=+29.428214469" Mar 6 02:37:36.665884 kubelet[2497]: I0306 02:37:36.665077 2497 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=8.6650632 podStartE2EDuration="8.6650632s" podCreationTimestamp="2026-03-06 02:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 02:37:29.974734464 +0000 UTC m=+29.483901506" watchObservedRunningTime="2026-03-06 02:37:36.6650632 +0000 UTC m=+36.174230232" Mar 6 02:37:36.768857 systemd[1]: Reload requested from client PID 2791 ('systemctl') (unit session-7.scope)... Mar 6 02:37:36.768890 systemd[1]: Reloading... Mar 6 02:37:37.468850 zram_generator::config[2837]: No configuration found. Mar 6 02:37:39.829993 systemd[1]: Reloading finished in 3025 ms. Mar 6 02:37:40.144234 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:37:40.416678 systemd[1]: kubelet.service: Deactivated successfully. Mar 6 02:37:40.418185 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:37:40.418680 systemd[1]: kubelet.service: Consumed 8.373s CPU time, 131.5M memory peak. Mar 6 02:37:40.535198 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 02:37:41.018965 update_engine[1551]: I20260306 02:37:40.985116 1551 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Mar 6 02:37:41.018965 update_engine[1551]: I20260306 02:37:41.013621 1551 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Mar 6 02:37:41.051146 update_engine[1551]: I20260306 02:37:41.049028 1551 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Mar 6 02:37:41.055826 update_engine[1551]: I20260306 02:37:41.053103 1551 omaha_request_params.cc:62] Current group set to stable Mar 6 02:37:41.055826 update_engine[1551]: I20260306 02:37:41.054746 1551 update_attempter.cc:499] Already updated boot flags. Skipping. Mar 6 02:37:41.055826 update_engine[1551]: I20260306 02:37:41.054768 1551 update_attempter.cc:643] Scheduling an action processor start. Mar 6 02:37:41.055826 update_engine[1551]: I20260306 02:37:41.054883 1551 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Mar 6 02:37:41.056813 update_engine[1551]: I20260306 02:37:41.056236 1551 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Mar 6 02:37:41.063564 update_engine[1551]: I20260306 02:37:41.061975 1551 omaha_request_action.cc:271] Posting an Omaha request to disabled Mar 6 02:37:41.067553 update_engine[1551]: I20260306 02:37:41.064620 1551 omaha_request_action.cc:272] Request: Mar 6 02:37:41.067553 update_engine[1551]: Mar 6 02:37:41.067553 update_engine[1551]: Mar 6 02:37:41.067553 update_engine[1551]: Mar 6 02:37:41.067553 update_engine[1551]: Mar 6 02:37:41.067553 update_engine[1551]: Mar 6 02:37:41.067553 update_engine[1551]: Mar 6 02:37:41.067553 update_engine[1551]: Mar 6 02:37:41.067553 update_engine[1551]: Mar 6 02:37:41.067553 update_engine[1551]: I20260306 02:37:41.064823 1551 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 6 02:37:41.139924 update_engine[1551]: I20260306 02:37:41.088980 1551 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 6 02:37:41.149232 locksmithd[1606]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Mar 6 02:37:41.171642 update_engine[1551]: I20260306 02:37:41.170214 1551 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 6 02:37:41.186778 update_engine[1551]: E20260306 02:37:41.186719 1551 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 6 02:37:41.187639 update_engine[1551]: I20260306 02:37:41.187603 1551 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Mar 6 02:37:43.310967 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 02:37:43.937948 (kubelet)[2879]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 6 02:37:45.159089 kubelet[2879]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 6 02:37:45.282589 kubelet[2879]: I0306 02:37:45.282090 2879 server.go:525] "Kubelet version" kubeletVersion="v1.35.1" Mar 6 02:37:45.282774 kubelet[2879]: I0306 02:37:45.282272 2879 server.go:527] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 6 02:37:45.282774 kubelet[2879]: I0306 02:37:45.282652 2879 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 6 02:37:45.282774 kubelet[2879]: I0306 02:37:45.282663 2879 watchdog_linux.go:138] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 6 02:37:45.283266 kubelet[2879]: I0306 02:37:45.283124 2879 server.go:951] "Client rotation is on, will bootstrap in background" Mar 6 02:37:45.303843 kubelet[2879]: I0306 02:37:45.303240 2879 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 6 02:37:45.331741 kubelet[2879]: I0306 02:37:45.331071 2879 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 6 02:37:45.429979 kubelet[2879]: I0306 02:37:45.429850 2879 server.go:1418] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 6 02:37:45.461711 kubelet[2879]: I0306 02:37:45.461578 2879 server.go:775] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 6 02:37:45.463543 kubelet[2879]: I0306 02:37:45.462843 2879 container_manager_linux.go:272] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 6 02:37:45.463543 kubelet[2879]: I0306 02:37:45.462884 2879 container_manager_linux.go:277] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 6 02:37:45.465586 kubelet[2879]: I0306 02:37:45.463895 2879 topology_manager.go:143] "Creating topology manager with none policy" Mar 6 02:37:45.465586 kubelet[2879]: I0306 02:37:45.463916 2879 container_manager_linux.go:308] "Creating device plugin manager" Mar 6 02:37:45.465586 kubelet[2879]: I0306 02:37:45.463949 2879 container_manager_linux.go:317] "Creating Dynamic Resource Allocation (DRA) manager" Mar 6 02:37:45.465586 kubelet[2879]: I0306 02:37:45.464192 2879 state_mem.go:41] "Initialized" logger="CPUManager state memory" Mar 6 02:37:45.467038 kubelet[2879]: I0306 02:37:45.466991 2879 kubelet.go:482] "Attempting to sync node with API server" Mar 6 02:37:45.467038 kubelet[2879]: I0306 02:37:45.467018 2879 kubelet.go:383] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 6 02:37:45.472917 kubelet[2879]: I0306 02:37:45.472809 2879 kubelet.go:394] "Adding apiserver pod source" Mar 6 02:37:45.472917 kubelet[2879]: I0306 02:37:45.472846 2879 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 6 02:37:45.485776 kubelet[2879]: I0306 02:37:45.485229 2879 kuberuntime_manager.go:294] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Mar 6 02:37:45.495602 kubelet[2879]: I0306 02:37:45.494133 2879 kubelet.go:943] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 6 02:37:45.509573 kubelet[2879]: I0306 02:37:45.509246 2879 kubelet.go:970] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 6 02:37:45.589235 kubelet[2879]: I0306 02:37:45.589101 2879 server.go:1257] "Started kubelet" Mar 6 02:37:45.610022 kubelet[2879]: I0306 02:37:45.592141 2879 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 6 02:37:45.649913 kubelet[2879]: I0306 02:37:45.617847 2879 server.go:182] "Starting to listen" address="0.0.0.0" port=10250 Mar 6 02:37:45.770821 kubelet[2879]: I0306 02:37:45.760828 2879 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 6 02:37:45.819905 kubelet[2879]: I0306 02:37:45.817864 2879 fs_resource_analyzer.go:69] "Starting FS ResourceAnalyzer" Mar 6 02:37:45.833183 kubelet[2879]: I0306 02:37:45.833147 2879 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 6 02:37:45.845210 kubelet[2879]: I0306 02:37:45.845180 2879 server.go:254] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 6 02:37:45.882649 kubelet[2879]: I0306 02:37:45.864722 2879 volume_manager.go:311] "Starting Kubelet Volume Manager" Mar 6 02:37:45.907611 kubelet[2879]: I0306 02:37:45.889869 2879 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 6 02:37:45.908102 kubelet[2879]: I0306 02:37:45.907909 2879 reconciler.go:29] "Reconciler: start to sync state" Mar 6 02:37:45.924591 kubelet[2879]: I0306 02:37:45.920171 2879 factory.go:223] Registration of the systemd container factory successfully Mar 6 02:37:45.926791 kubelet[2879]: I0306 02:37:45.926759 2879 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 6 02:37:45.974239 kubelet[2879]: I0306 02:37:45.973834 2879 factory.go:223] Registration of the containerd container factory successfully Mar 6 02:37:46.033274 kubelet[2879]: I0306 02:37:46.029836 2879 server.go:317] "Adding debug handlers to kubelet server" Mar 6 02:37:46.258704 kubelet[2879]: I0306 02:37:46.257036 2879 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 6 02:37:46.267702 kubelet[2879]: I0306 02:37:46.267595 2879 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 6 02:37:46.267702 kubelet[2879]: I0306 02:37:46.267693 2879 status_manager.go:249] "Starting to sync pod status with apiserver" Mar 6 02:37:46.267894 kubelet[2879]: I0306 02:37:46.267731 2879 kubelet.go:2501] "Starting kubelet main sync loop" Mar 6 02:37:46.267934 kubelet[2879]: E0306 02:37:46.267896 2879 kubelet.go:2525] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 6 02:37:46.556752 kubelet[2879]: E0306 02:37:46.416838 2879 kubelet.go:2525] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 6 02:37:46.556752 kubelet[2879]: I0306 02:37:46.506163 2879 apiserver.go:52] "Watching apiserver" Mar 6 02:37:46.628139 kubelet[2879]: E0306 02:37:46.623059 2879 kubelet.go:2525] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 6 02:37:46.722105 kubelet[2879]: I0306 02:37:46.722021 2879 cpu_manager.go:225] "Starting" policy="none" Mar 6 02:37:46.722105 kubelet[2879]: I0306 02:37:46.722096 2879 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 6 02:37:46.722846 kubelet[2879]: I0306 02:37:46.722124 2879 state_mem.go:41] "Initialized" logger="CPUManager state checkpoint.CPUManager state memory" Mar 6 02:37:46.722846 kubelet[2879]: I0306 02:37:46.722518 2879 state_mem.go:94] "Updated default CPUSet" logger="CPUManager state checkpoint.CPUManager state memory" cpuSet="" Mar 6 02:37:46.722846 kubelet[2879]: I0306 02:37:46.722539 2879 state_mem.go:102] "Updated CPUSet assignments" logger="CPUManager state checkpoint.CPUManager state memory" assignments={} Mar 6 02:37:46.722846 kubelet[2879]: I0306 02:37:46.722572 2879 policy_none.go:50] "Start" Mar 6 02:37:46.722846 kubelet[2879]: I0306 02:37:46.722585 2879 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 6 02:37:46.722846 kubelet[2879]: I0306 02:37:46.722601 2879 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 6 02:37:46.722994 kubelet[2879]: I0306 02:37:46.722905 2879 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Mar 6 02:37:46.722994 kubelet[2879]: I0306 02:37:46.722923 2879 policy_none.go:44] "Start" Mar 6 02:37:46.746172 kubelet[2879]: E0306 02:37:46.745837 2879 manager.go:525] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 6 02:37:46.746624 kubelet[2879]: I0306 02:37:46.746538 2879 eviction_manager.go:194] "Eviction manager: starting control loop" Mar 6 02:37:46.746700 kubelet[2879]: I0306 02:37:46.746612 2879 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 6 02:37:46.755850 kubelet[2879]: I0306 02:37:46.750019 2879 plugin_manager.go:121] "Starting Kubelet Plugin Manager" Mar 6 02:37:46.755850 kubelet[2879]: I0306 02:37:46.752686 2879 kuberuntime_manager.go:2062] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 6 02:37:46.755850 kubelet[2879]: E0306 02:37:46.753479 2879 eviction_manager.go:272] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 6 02:37:46.756007 containerd[1569]: time="2026-03-06T02:37:46.755568017Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 6 02:37:46.763848 kubelet[2879]: I0306 02:37:46.759788 2879 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 6 02:37:47.046635 kubelet[2879]: I0306 02:37:47.035909 2879 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Mar 6 02:37:47.083531 kubelet[2879]: I0306 02:37:47.082379 2879 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 6 02:37:47.088518 kubelet[2879]: I0306 02:37:47.087166 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bd81bb6a14e176da833e3a8030ee5eac-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"bd81bb6a14e176da833e3a8030ee5eac\") " pod="kube-system/kube-scheduler-localhost" Mar 6 02:37:47.143854 kubelet[2879]: E0306 02:37:47.141880 2879 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Mar 6 02:37:47.151058 systemd[1]: Created slice kubepods-besteffort-podf33f8c42_c991_45cb_a1d7_d9f1e8c25012.slice - libcontainer container kubepods-besteffort-podf33f8c42_c991_45cb_a1d7_d9f1e8c25012.slice. Mar 6 02:37:47.180417 kubelet[2879]: I0306 02:37:47.179962 2879 kubelet_node_status.go:123] "Node was previously registered" node="localhost" Mar 6 02:37:47.180417 kubelet[2879]: I0306 02:37:47.180056 2879 kubelet_node_status.go:77] "Successfully registered node" node="localhost" Mar 6 02:37:47.189143 kubelet[2879]: I0306 02:37:47.189109 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7e99612e9106acc372b437b601cabe3e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"7e99612e9106acc372b437b601cabe3e\") " pod="kube-system/kube-apiserver-localhost" Mar 6 02:37:47.199543 kubelet[2879]: I0306 02:37:47.189400 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7e99612e9106acc372b437b601cabe3e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"7e99612e9106acc372b437b601cabe3e\") " pod="kube-system/kube-apiserver-localhost" Mar 6 02:37:47.199543 kubelet[2879]: I0306 02:37:47.191659 2879 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 6 02:37:47.199543 kubelet[2879]: I0306 02:37:47.192404 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7e99612e9106acc372b437b601cabe3e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"7e99612e9106acc372b437b601cabe3e\") " pod="kube-system/kube-apiserver-localhost" Mar 6 02:37:47.199543 kubelet[2879]: I0306 02:37:47.192604 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:47.199543 kubelet[2879]: I0306 02:37:47.192631 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:47.199543 kubelet[2879]: I0306 02:37:47.192650 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:47.200142 kubelet[2879]: I0306 02:37:47.192675 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:47.200142 kubelet[2879]: I0306 02:37:47.192697 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/f33f8c42-c991-45cb-a1d7-d9f1e8c25012-kube-proxy\") pod \"kube-proxy-6424l\" (UID: \"f33f8c42-c991-45cb-a1d7-d9f1e8c25012\") " pod="kube-system/kube-proxy-6424l" Mar 6 02:37:47.200142 kubelet[2879]: I0306 02:37:47.192875 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f420dd303687d038b2bc2fa1d277c55c-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"f420dd303687d038b2bc2fa1d277c55c\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 02:37:47.200142 kubelet[2879]: I0306 02:37:47.192910 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f33f8c42-c991-45cb-a1d7-d9f1e8c25012-xtables-lock\") pod \"kube-proxy-6424l\" (UID: \"f33f8c42-c991-45cb-a1d7-d9f1e8c25012\") " pod="kube-system/kube-proxy-6424l" Mar 6 02:37:47.200142 kubelet[2879]: I0306 02:37:47.192935 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f33f8c42-c991-45cb-a1d7-d9f1e8c25012-lib-modules\") pod \"kube-proxy-6424l\" (UID: \"f33f8c42-c991-45cb-a1d7-d9f1e8c25012\") " pod="kube-system/kube-proxy-6424l" Mar 6 02:37:47.201738 kubelet[2879]: I0306 02:37:47.192956 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fvd8\" (UniqueName: \"kubernetes.io/projected/f33f8c42-c991-45cb-a1d7-d9f1e8c25012-kube-api-access-8fvd8\") pod \"kube-proxy-6424l\" (UID: \"f33f8c42-c991-45cb-a1d7-d9f1e8c25012\") " pod="kube-system/kube-proxy-6424l" Mar 6 02:37:48.228490 containerd[1569]: time="2026-03-06T02:37:48.226228846Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-6424l,Uid:f33f8c42-c991-45cb-a1d7-d9f1e8c25012,Namespace:kube-system,Attempt:0,}" Mar 6 02:37:48.692655 containerd[1569]: time="2026-03-06T02:37:48.692115512Z" level=info msg="connecting to shim f9f1918d5f7496af7fa8f5e9033bc64e230d32a5cced6be2ba84745747c69b84" address="unix:///run/containerd/s/226270ffce2d4ff950b5a367f103ea09d39f9fdd76dfc56588eecc24abf095b0" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:37:49.735955 systemd[1]: Started cri-containerd-f9f1918d5f7496af7fa8f5e9033bc64e230d32a5cced6be2ba84745747c69b84.scope - libcontainer container f9f1918d5f7496af7fa8f5e9033bc64e230d32a5cced6be2ba84745747c69b84. Mar 6 02:37:51.268762 containerd[1569]: time="2026-03-06T02:37:51.267102968Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-6424l,Uid:f33f8c42-c991-45cb-a1d7-d9f1e8c25012,Namespace:kube-system,Attempt:0,} returns sandbox id \"f9f1918d5f7496af7fa8f5e9033bc64e230d32a5cced6be2ba84745747c69b84\"" Mar 6 02:37:51.395620 containerd[1569]: time="2026-03-06T02:37:51.394951183Z" level=info msg="CreateContainer within sandbox \"f9f1918d5f7496af7fa8f5e9033bc64e230d32a5cced6be2ba84745747c69b84\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 6 02:37:51.751273 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1196452720.mount: Deactivated successfully. Mar 6 02:37:51.767523 containerd[1569]: time="2026-03-06T02:37:51.767176375Z" level=info msg="Container 7fe83c03c35706959b547737f541e4ee1dd473642c25f754b120ee86b4bda0a9: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:37:51.824677 containerd[1569]: time="2026-03-06T02:37:51.824515466Z" level=info msg="CreateContainer within sandbox \"f9f1918d5f7496af7fa8f5e9033bc64e230d32a5cced6be2ba84745747c69b84\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"7fe83c03c35706959b547737f541e4ee1dd473642c25f754b120ee86b4bda0a9\"" Mar 6 02:37:51.825960 containerd[1569]: time="2026-03-06T02:37:51.825927920Z" level=info msg="StartContainer for \"7fe83c03c35706959b547737f541e4ee1dd473642c25f754b120ee86b4bda0a9\"" Mar 6 02:37:51.829682 containerd[1569]: time="2026-03-06T02:37:51.829648328Z" level=info msg="connecting to shim 7fe83c03c35706959b547737f541e4ee1dd473642c25f754b120ee86b4bda0a9" address="unix:///run/containerd/s/226270ffce2d4ff950b5a367f103ea09d39f9fdd76dfc56588eecc24abf095b0" protocol=ttrpc version=3 Mar 6 02:37:51.932899 update_engine[1551]: I20260306 02:37:51.930593 1551 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 6 02:37:51.932899 update_engine[1551]: I20260306 02:37:51.932191 1551 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 6 02:37:51.936091 update_engine[1551]: I20260306 02:37:51.933656 1551 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 6 02:37:51.959957 update_engine[1551]: E20260306 02:37:51.959172 1551 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 6 02:37:51.959957 update_engine[1551]: I20260306 02:37:51.959753 1551 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Mar 6 02:37:52.157245 systemd[1]: Started cri-containerd-7fe83c03c35706959b547737f541e4ee1dd473642c25f754b120ee86b4bda0a9.scope - libcontainer container 7fe83c03c35706959b547737f541e4ee1dd473642c25f754b120ee86b4bda0a9. Mar 6 02:37:53.596697 containerd[1569]: time="2026-03-06T02:37:53.595919464Z" level=info msg="StartContainer for \"7fe83c03c35706959b547737f541e4ee1dd473642c25f754b120ee86b4bda0a9\" returns successfully" Mar 6 02:37:54.986913 kubelet[2879]: I0306 02:37:54.986262 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-proxy-6424l" podStartSLOduration=9.986240255 podStartE2EDuration="9.986240255s" podCreationTimestamp="2026-03-06 02:37:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 02:37:54.244984501 +0000 UTC m=+10.252183024" watchObservedRunningTime="2026-03-06 02:37:54.986240255 +0000 UTC m=+10.993438757" Mar 6 02:37:55.064808 systemd[1]: Created slice kubepods-besteffort-podc05c647f_627e_465b_b960_d6cfd77cacd2.slice - libcontainer container kubepods-besteffort-podc05c647f_627e_465b_b960_d6cfd77cacd2.slice. Mar 6 02:37:55.195829 kubelet[2879]: I0306 02:37:55.190954 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tjz8\" (UniqueName: \"kubernetes.io/projected/c05c647f-627e-465b-b960-d6cfd77cacd2-kube-api-access-8tjz8\") pod \"tigera-operator-6cf4cccc57-9l9tm\" (UID: \"c05c647f-627e-465b-b960-d6cfd77cacd2\") " pod="tigera-operator/tigera-operator-6cf4cccc57-9l9tm" Mar 6 02:37:55.195829 kubelet[2879]: I0306 02:37:55.193432 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/c05c647f-627e-465b-b960-d6cfd77cacd2-var-lib-calico\") pod \"tigera-operator-6cf4cccc57-9l9tm\" (UID: \"c05c647f-627e-465b-b960-d6cfd77cacd2\") " pod="tigera-operator/tigera-operator-6cf4cccc57-9l9tm" Mar 6 02:37:55.645092 containerd[1569]: time="2026-03-06T02:37:55.643046669Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6cf4cccc57-9l9tm,Uid:c05c647f-627e-465b-b960-d6cfd77cacd2,Namespace:tigera-operator,Attempt:0,}" Mar 6 02:37:55.921630 containerd[1569]: time="2026-03-06T02:37:55.900104320Z" level=info msg="connecting to shim 6a689746ba7dc6524589b1f1fd2ea02fca5244d1739940fe878f1e3911b24a6c" address="unix:///run/containerd/s/6ba11aff529c4fdd2ffc204f47a933961786f0add67b7481f462f8a6332d565d" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:37:56.811165 systemd[1]: Started cri-containerd-6a689746ba7dc6524589b1f1fd2ea02fca5244d1739940fe878f1e3911b24a6c.scope - libcontainer container 6a689746ba7dc6524589b1f1fd2ea02fca5244d1739940fe878f1e3911b24a6c. Mar 6 02:37:57.495245 containerd[1569]: time="2026-03-06T02:37:57.495005129Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6cf4cccc57-9l9tm,Uid:c05c647f-627e-465b-b960-d6cfd77cacd2,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"6a689746ba7dc6524589b1f1fd2ea02fca5244d1739940fe878f1e3911b24a6c\"" Mar 6 02:37:57.506050 containerd[1569]: time="2026-03-06T02:37:57.504950230Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\"" Mar 6 02:37:59.397533 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount98030403.mount: Deactivated successfully. Mar 6 02:38:01.939093 update_engine[1551]: I20260306 02:38:01.932009 1551 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 6 02:38:01.939093 update_engine[1551]: I20260306 02:38:01.933557 1551 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 6 02:38:01.939093 update_engine[1551]: I20260306 02:38:01.934654 1551 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 6 02:38:01.951813 update_engine[1551]: E20260306 02:38:01.951582 1551 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 6 02:38:01.951813 update_engine[1551]: I20260306 02:38:01.951749 1551 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Mar 6 02:38:11.932994 update_engine[1551]: I20260306 02:38:11.928428 1551 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 6 02:38:11.932994 update_engine[1551]: I20260306 02:38:11.930847 1551 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 6 02:38:11.935937 update_engine[1551]: I20260306 02:38:11.933929 1551 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 6 02:38:12.158128 update_engine[1551]: E20260306 02:38:12.156386 1551 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 6 02:38:12.158128 update_engine[1551]: I20260306 02:38:12.156798 1551 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Mar 6 02:38:12.158128 update_engine[1551]: I20260306 02:38:12.156816 1551 omaha_request_action.cc:617] Omaha request response: Mar 6 02:38:12.160845 update_engine[1551]: E20260306 02:38:12.159080 1551 omaha_request_action.cc:636] Omaha request network transfer failed. Mar 6 02:38:12.160845 update_engine[1551]: I20260306 02:38:12.159641 1551 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Mar 6 02:38:12.160845 update_engine[1551]: I20260306 02:38:12.159657 1551 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 6 02:38:12.160845 update_engine[1551]: I20260306 02:38:12.159667 1551 update_attempter.cc:306] Processing Done. Mar 6 02:38:12.160845 update_engine[1551]: E20260306 02:38:12.159823 1551 update_attempter.cc:619] Update failed. Mar 6 02:38:12.160845 update_engine[1551]: I20260306 02:38:12.159884 1551 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Mar 6 02:38:12.160845 update_engine[1551]: I20260306 02:38:12.159899 1551 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Mar 6 02:38:12.160845 update_engine[1551]: I20260306 02:38:12.159909 1551 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Mar 6 02:38:12.161174 update_engine[1551]: I20260306 02:38:12.161149 1551 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Mar 6 02:38:12.161543 update_engine[1551]: I20260306 02:38:12.161271 1551 omaha_request_action.cc:271] Posting an Omaha request to disabled Mar 6 02:38:12.161633 update_engine[1551]: I20260306 02:38:12.161612 1551 omaha_request_action.cc:272] Request: Mar 6 02:38:12.161633 update_engine[1551]: Mar 6 02:38:12.161633 update_engine[1551]: Mar 6 02:38:12.161633 update_engine[1551]: Mar 6 02:38:12.161633 update_engine[1551]: Mar 6 02:38:12.161633 update_engine[1551]: Mar 6 02:38:12.161633 update_engine[1551]: Mar 6 02:38:12.161862 update_engine[1551]: I20260306 02:38:12.161841 1551 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 6 02:38:12.161958 update_engine[1551]: I20260306 02:38:12.161937 1551 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 6 02:38:12.162828 update_engine[1551]: I20260306 02:38:12.162778 1551 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 6 02:38:12.166572 locksmithd[1606]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Mar 6 02:38:12.181868 update_engine[1551]: E20260306 02:38:12.181742 1551 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Mar 6 02:38:12.182224 update_engine[1551]: I20260306 02:38:12.181901 1551 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Mar 6 02:38:12.182224 update_engine[1551]: I20260306 02:38:12.181918 1551 omaha_request_action.cc:617] Omaha request response: Mar 6 02:38:12.182224 update_engine[1551]: I20260306 02:38:12.181931 1551 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 6 02:38:12.182224 update_engine[1551]: I20260306 02:38:12.181941 1551 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 6 02:38:12.182224 update_engine[1551]: I20260306 02:38:12.181949 1551 update_attempter.cc:306] Processing Done. Mar 6 02:38:12.182224 update_engine[1551]: I20260306 02:38:12.181962 1551 update_attempter.cc:310] Error event sent. Mar 6 02:38:12.182224 update_engine[1551]: I20260306 02:38:12.181982 1551 update_check_scheduler.cc:74] Next update check in 42m38s Mar 6 02:38:12.188077 locksmithd[1606]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Mar 6 02:38:13.899365 containerd[1569]: time="2026-03-06T02:38:13.899102318Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:13.903130 containerd[1569]: time="2026-03-06T02:38:13.903050121Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.7: active requests=0, bytes read=40846156" Mar 6 02:38:13.908652 containerd[1569]: time="2026-03-06T02:38:13.908537678Z" level=info msg="ImageCreate event name:\"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:13.915512 containerd[1569]: time="2026-03-06T02:38:13.915398100Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:13.917245 containerd[1569]: time="2026-03-06T02:38:13.917051630Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.7\" with image id \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\", repo tag \"quay.io/tigera/operator:v1.40.7\", repo digest \"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\", size \"40842151\" in 16.411767105s" Mar 6 02:38:13.917245 containerd[1569]: time="2026-03-06T02:38:13.917153340Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\" returns image reference \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\"" Mar 6 02:38:13.931849 containerd[1569]: time="2026-03-06T02:38:13.931400368Z" level=info msg="CreateContainer within sandbox \"6a689746ba7dc6524589b1f1fd2ea02fca5244d1739940fe878f1e3911b24a6c\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Mar 6 02:38:13.966767 containerd[1569]: time="2026-03-06T02:38:13.965752630Z" level=info msg="Container 8d38c688c316929de2899c98d24bf9b37c8463e710d8119f82a14b17a83d5278: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:38:14.084789 containerd[1569]: time="2026-03-06T02:38:14.080610454Z" level=info msg="CreateContainer within sandbox \"6a689746ba7dc6524589b1f1fd2ea02fca5244d1739940fe878f1e3911b24a6c\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"8d38c688c316929de2899c98d24bf9b37c8463e710d8119f82a14b17a83d5278\"" Mar 6 02:38:14.096549 containerd[1569]: time="2026-03-06T02:38:14.090154396Z" level=info msg="StartContainer for \"8d38c688c316929de2899c98d24bf9b37c8463e710d8119f82a14b17a83d5278\"" Mar 6 02:38:14.096549 containerd[1569]: time="2026-03-06T02:38:14.095752051Z" level=info msg="connecting to shim 8d38c688c316929de2899c98d24bf9b37c8463e710d8119f82a14b17a83d5278" address="unix:///run/containerd/s/6ba11aff529c4fdd2ffc204f47a933961786f0add67b7481f462f8a6332d565d" protocol=ttrpc version=3 Mar 6 02:38:14.327990 systemd[1]: Started cri-containerd-8d38c688c316929de2899c98d24bf9b37c8463e710d8119f82a14b17a83d5278.scope - libcontainer container 8d38c688c316929de2899c98d24bf9b37c8463e710d8119f82a14b17a83d5278. Mar 6 02:38:14.824844 containerd[1569]: time="2026-03-06T02:38:14.818273723Z" level=info msg="StartContainer for \"8d38c688c316929de2899c98d24bf9b37c8463e710d8119f82a14b17a83d5278\" returns successfully" Mar 6 02:38:27.377407 sudo[1787]: pam_unix(sudo:session): session closed for user root Mar 6 02:38:27.387759 sshd[1786]: Connection closed by 10.0.0.1 port 38322 Mar 6 02:38:27.396645 sshd-session[1783]: pam_unix(sshd:session): session closed for user core Mar 6 02:38:27.418458 systemd-logind[1544]: Session 7 logged out. Waiting for processes to exit. Mar 6 02:38:27.418935 systemd[1]: sshd@6-10.0.0.69:22-10.0.0.1:38322.service: Deactivated successfully. Mar 6 02:38:27.426694 systemd[1]: session-7.scope: Deactivated successfully. Mar 6 02:38:27.428673 systemd[1]: session-7.scope: Consumed 24.068s CPU time, 230.2M memory peak. Mar 6 02:38:27.442868 systemd-logind[1544]: Removed session 7. Mar 6 02:38:34.520730 kubelet[2879]: I0306 02:38:34.520056 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6cf4cccc57-9l9tm" podStartSLOduration=24.102393301 podStartE2EDuration="40.520014992s" podCreationTimestamp="2026-03-06 02:37:54 +0000 UTC" firstStartedPulling="2026-03-06 02:37:57.502109844 +0000 UTC m=+13.509308336" lastFinishedPulling="2026-03-06 02:38:13.919731535 +0000 UTC m=+29.926930027" observedRunningTime="2026-03-06 02:38:15.966038381 +0000 UTC m=+31.973237104" watchObservedRunningTime="2026-03-06 02:38:34.520014992 +0000 UTC m=+50.527213484" Mar 6 02:38:34.565547 systemd[1]: Created slice kubepods-besteffort-pod005aac40_e789_4801_bb37_c5b65b298482.slice - libcontainer container kubepods-besteffort-pod005aac40_e789_4801_bb37_c5b65b298482.slice. Mar 6 02:38:34.574407 kubelet[2879]: I0306 02:38:34.574116 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/005aac40-e789-4801-bb37-c5b65b298482-tigera-ca-bundle\") pod \"calico-typha-75c59bd9f9-vcbwp\" (UID: \"005aac40-e789-4801-bb37-c5b65b298482\") " pod="calico-system/calico-typha-75c59bd9f9-vcbwp" Mar 6 02:38:34.575827 kubelet[2879]: I0306 02:38:34.575695 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/005aac40-e789-4801-bb37-c5b65b298482-typha-certs\") pod \"calico-typha-75c59bd9f9-vcbwp\" (UID: \"005aac40-e789-4801-bb37-c5b65b298482\") " pod="calico-system/calico-typha-75c59bd9f9-vcbwp" Mar 6 02:38:34.575827 kubelet[2879]: I0306 02:38:34.575819 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccsrj\" (UniqueName: \"kubernetes.io/projected/005aac40-e789-4801-bb37-c5b65b298482-kube-api-access-ccsrj\") pod \"calico-typha-75c59bd9f9-vcbwp\" (UID: \"005aac40-e789-4801-bb37-c5b65b298482\") " pod="calico-system/calico-typha-75c59bd9f9-vcbwp" Mar 6 02:38:34.898442 systemd[1]: Created slice kubepods-besteffort-poded53f349_d1f4_4f37_816e_cefd154747f5.slice - libcontainer container kubepods-besteffort-poded53f349_d1f4_4f37_816e_cefd154747f5.slice. Mar 6 02:38:34.912445 containerd[1569]: time="2026-03-06T02:38:34.911746446Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-75c59bd9f9-vcbwp,Uid:005aac40-e789-4801-bb37-c5b65b298482,Namespace:calico-system,Attempt:0,}" Mar 6 02:38:34.988451 kubelet[2879]: I0306 02:38:34.981545 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/ed53f349-d1f4-4f37-816e-cefd154747f5-node-certs\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.988451 kubelet[2879]: I0306 02:38:34.981596 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-var-run-calico\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.988451 kubelet[2879]: I0306 02:38:34.981713 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-cni-log-dir\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.988451 kubelet[2879]: I0306 02:38:34.981735 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-nodeproc\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.988451 kubelet[2879]: I0306 02:38:34.981756 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-policysync\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.992938 kubelet[2879]: I0306 02:38:34.981776 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-sys-fs\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.992938 kubelet[2879]: I0306 02:38:34.981793 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed53f349-d1f4-4f37-816e-cefd154747f5-tigera-ca-bundle\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.992938 kubelet[2879]: I0306 02:38:34.981821 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-cni-bin-dir\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.992938 kubelet[2879]: I0306 02:38:34.981842 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-flexvol-driver-host\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.992938 kubelet[2879]: I0306 02:38:34.981861 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-var-lib-calico\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.993500 kubelet[2879]: I0306 02:38:34.981878 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-xtables-lock\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.993500 kubelet[2879]: I0306 02:38:34.981897 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bkg8\" (UniqueName: \"kubernetes.io/projected/ed53f349-d1f4-4f37-816e-cefd154747f5-kube-api-access-9bkg8\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.993500 kubelet[2879]: I0306 02:38:34.981929 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-lib-modules\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.993500 kubelet[2879]: I0306 02:38:34.981950 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-cni-net-dir\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:34.993500 kubelet[2879]: I0306 02:38:34.981974 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/ed53f349-d1f4-4f37-816e-cefd154747f5-bpffs\") pod \"calico-node-cvqkx\" (UID: \"ed53f349-d1f4-4f37-816e-cefd154747f5\") " pod="calico-system/calico-node-cvqkx" Mar 6 02:38:35.104098 kubelet[2879]: E0306 02:38:35.103899 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.104098 kubelet[2879]: W0306 02:38:35.103929 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.104098 kubelet[2879]: E0306 02:38:35.104043 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.125209 kubelet[2879]: E0306 02:38:35.125175 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.127396 kubelet[2879]: W0306 02:38:35.127137 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.128796 kubelet[2879]: E0306 02:38:35.128722 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.182498 containerd[1569]: time="2026-03-06T02:38:35.180004196Z" level=info msg="connecting to shim 9a8cae8d13218009569f1f0a48184558c37ee12e13a4fd8c8f0fd84a3b6ba46e" address="unix:///run/containerd/s/c80a726a51b8cbf181088b40524d32551a2147593411873bf18208aab85458fc" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:38:35.211272 kubelet[2879]: E0306 02:38:35.205175 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:35.247552 kubelet[2879]: E0306 02:38:35.247515 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.247847 kubelet[2879]: W0306 02:38:35.247822 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.247955 kubelet[2879]: E0306 02:38:35.247936 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.259771 kubelet[2879]: E0306 02:38:35.259742 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.259926 kubelet[2879]: W0306 02:38:35.259904 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.260019 kubelet[2879]: E0306 02:38:35.260002 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.263557 kubelet[2879]: E0306 02:38:35.263538 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.263773 kubelet[2879]: W0306 02:38:35.263751 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.263878 kubelet[2879]: E0306 02:38:35.263860 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.264493 kubelet[2879]: E0306 02:38:35.264473 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.264800 kubelet[2879]: W0306 02:38:35.264781 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.264886 kubelet[2879]: E0306 02:38:35.264870 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.267115 kubelet[2879]: E0306 02:38:35.267098 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.267516 kubelet[2879]: W0306 02:38:35.267492 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.267700 kubelet[2879]: E0306 02:38:35.267593 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.270242 kubelet[2879]: E0306 02:38:35.270224 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.271101 kubelet[2879]: W0306 02:38:35.271077 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.271575 kubelet[2879]: E0306 02:38:35.271168 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.274894 kubelet[2879]: E0306 02:38:35.274711 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.275014 kubelet[2879]: W0306 02:38:35.274974 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.275772 kubelet[2879]: E0306 02:38:35.275751 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.277217 kubelet[2879]: E0306 02:38:35.277141 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.277217 kubelet[2879]: W0306 02:38:35.277158 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.277217 kubelet[2879]: E0306 02:38:35.277175 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.286504 kubelet[2879]: E0306 02:38:35.285825 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.286504 kubelet[2879]: W0306 02:38:35.285844 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.286504 kubelet[2879]: E0306 02:38:35.285863 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.290508 kubelet[2879]: E0306 02:38:35.287492 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.290508 kubelet[2879]: W0306 02:38:35.287504 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.290508 kubelet[2879]: E0306 02:38:35.287520 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.291575 kubelet[2879]: E0306 02:38:35.291550 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.291985 kubelet[2879]: W0306 02:38:35.291749 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.291985 kubelet[2879]: E0306 02:38:35.291780 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.297550 kubelet[2879]: E0306 02:38:35.297529 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.298391 kubelet[2879]: W0306 02:38:35.298366 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.312918 kubelet[2879]: E0306 02:38:35.312894 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.320139 kubelet[2879]: E0306 02:38:35.320117 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.320252 kubelet[2879]: W0306 02:38:35.320234 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.320722 kubelet[2879]: E0306 02:38:35.320524 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.339003 kubelet[2879]: E0306 02:38:35.338976 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.339176 kubelet[2879]: W0306 02:38:35.339156 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.339268 kubelet[2879]: E0306 02:38:35.339251 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.343564 kubelet[2879]: E0306 02:38:35.340440 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.343564 kubelet[2879]: W0306 02:38:35.340460 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.343564 kubelet[2879]: E0306 02:38:35.340478 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.347880 kubelet[2879]: E0306 02:38:35.347518 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.348153 kubelet[2879]: W0306 02:38:35.348131 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.348547 kubelet[2879]: E0306 02:38:35.348526 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.349979 kubelet[2879]: E0306 02:38:35.349962 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.350058 kubelet[2879]: W0306 02:38:35.350043 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.350783 kubelet[2879]: E0306 02:38:35.350475 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.354999 kubelet[2879]: E0306 02:38:35.354981 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.356486 kubelet[2879]: W0306 02:38:35.355269 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.356592 kubelet[2879]: E0306 02:38:35.356575 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.363235 kubelet[2879]: E0306 02:38:35.362928 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.364037 kubelet[2879]: W0306 02:38:35.364018 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.364109 kubelet[2879]: E0306 02:38:35.364097 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.367092 kubelet[2879]: E0306 02:38:35.365507 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.367199 kubelet[2879]: W0306 02:38:35.367180 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.367273 kubelet[2879]: E0306 02:38:35.367259 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.375143 kubelet[2879]: E0306 02:38:35.375120 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.375542 kubelet[2879]: W0306 02:38:35.375518 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.375713 kubelet[2879]: E0306 02:38:35.375696 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.393773 kubelet[2879]: E0306 02:38:35.393740 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.394778 kubelet[2879]: W0306 02:38:35.394463 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.394922 kubelet[2879]: E0306 02:38:35.394900 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.398230 kubelet[2879]: I0306 02:38:35.398204 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/74be49b3-6c58-4ae9-86a2-c8139f48d9c9-kubelet-dir\") pod \"csi-node-driver-gqfqd\" (UID: \"74be49b3-6c58-4ae9-86a2-c8139f48d9c9\") " pod="calico-system/csi-node-driver-gqfqd" Mar 6 02:38:35.400757 kubelet[2879]: E0306 02:38:35.400596 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.400757 kubelet[2879]: W0306 02:38:35.400715 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.400757 kubelet[2879]: E0306 02:38:35.400737 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.400947 kubelet[2879]: I0306 02:38:35.400925 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gwg6\" (UniqueName: \"kubernetes.io/projected/74be49b3-6c58-4ae9-86a2-c8139f48d9c9-kube-api-access-6gwg6\") pod \"csi-node-driver-gqfqd\" (UID: \"74be49b3-6c58-4ae9-86a2-c8139f48d9c9\") " pod="calico-system/csi-node-driver-gqfqd" Mar 6 02:38:35.420722 kubelet[2879]: E0306 02:38:35.420181 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.420722 kubelet[2879]: W0306 02:38:35.420199 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.420722 kubelet[2879]: E0306 02:38:35.420218 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.427951 kubelet[2879]: E0306 02:38:35.421855 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.427951 kubelet[2879]: W0306 02:38:35.421872 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.427951 kubelet[2879]: E0306 02:38:35.421888 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.430103 kubelet[2879]: E0306 02:38:35.429247 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.430103 kubelet[2879]: W0306 02:38:35.429269 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.430103 kubelet[2879]: E0306 02:38:35.429456 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.437482 kubelet[2879]: E0306 02:38:35.433106 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.437482 kubelet[2879]: W0306 02:38:35.433123 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.437482 kubelet[2879]: E0306 02:38:35.433140 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.437482 kubelet[2879]: I0306 02:38:35.433178 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/74be49b3-6c58-4ae9-86a2-c8139f48d9c9-registration-dir\") pod \"csi-node-driver-gqfqd\" (UID: \"74be49b3-6c58-4ae9-86a2-c8139f48d9c9\") " pod="calico-system/csi-node-driver-gqfqd" Mar 6 02:38:35.443846 kubelet[2879]: E0306 02:38:35.437850 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.443846 kubelet[2879]: W0306 02:38:35.437869 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.443846 kubelet[2879]: E0306 02:38:35.438961 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.443846 kubelet[2879]: E0306 02:38:35.443540 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.443846 kubelet[2879]: W0306 02:38:35.443568 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.443846 kubelet[2879]: E0306 02:38:35.443603 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.446164 kubelet[2879]: E0306 02:38:35.445597 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.446164 kubelet[2879]: W0306 02:38:35.445732 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.446164 kubelet[2879]: E0306 02:38:35.446139 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.446963 kubelet[2879]: I0306 02:38:35.446941 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/74be49b3-6c58-4ae9-86a2-c8139f48d9c9-socket-dir\") pod \"csi-node-driver-gqfqd\" (UID: \"74be49b3-6c58-4ae9-86a2-c8139f48d9c9\") " pod="calico-system/csi-node-driver-gqfqd" Mar 6 02:38:35.448496 kubelet[2879]: E0306 02:38:35.448456 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.448908 kubelet[2879]: W0306 02:38:35.448874 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.449003 kubelet[2879]: E0306 02:38:35.448986 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.451260 kubelet[2879]: E0306 02:38:35.451236 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.451697 kubelet[2879]: W0306 02:38:35.451582 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.451833 kubelet[2879]: E0306 02:38:35.451806 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.453151 kubelet[2879]: I0306 02:38:35.453125 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/74be49b3-6c58-4ae9-86a2-c8139f48d9c9-varrun\") pod \"csi-node-driver-gqfqd\" (UID: \"74be49b3-6c58-4ae9-86a2-c8139f48d9c9\") " pod="calico-system/csi-node-driver-gqfqd" Mar 6 02:38:35.463483 kubelet[2879]: E0306 02:38:35.463272 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.463699 kubelet[2879]: W0306 02:38:35.463585 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.463801 kubelet[2879]: E0306 02:38:35.463780 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.466006 kubelet[2879]: E0306 02:38:35.465983 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.466111 kubelet[2879]: W0306 02:38:35.466093 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.466200 kubelet[2879]: E0306 02:38:35.466182 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.474167 kubelet[2879]: E0306 02:38:35.474148 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.474527 kubelet[2879]: W0306 02:38:35.474264 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.474726 kubelet[2879]: E0306 02:38:35.474702 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.477189 kubelet[2879]: E0306 02:38:35.477173 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.480042 kubelet[2879]: W0306 02:38:35.480022 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.480242 kubelet[2879]: E0306 02:38:35.480225 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.493507 systemd[1]: Started cri-containerd-9a8cae8d13218009569f1f0a48184558c37ee12e13a4fd8c8f0fd84a3b6ba46e.scope - libcontainer container 9a8cae8d13218009569f1f0a48184558c37ee12e13a4fd8c8f0fd84a3b6ba46e. Mar 6 02:38:35.558102 containerd[1569]: time="2026-03-06T02:38:35.557534546Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-cvqkx,Uid:ed53f349-d1f4-4f37-816e-cefd154747f5,Namespace:calico-system,Attempt:0,}" Mar 6 02:38:35.560264 kubelet[2879]: E0306 02:38:35.559542 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.560264 kubelet[2879]: W0306 02:38:35.559570 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.560264 kubelet[2879]: E0306 02:38:35.559595 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.561827 kubelet[2879]: E0306 02:38:35.560978 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.561827 kubelet[2879]: W0306 02:38:35.560994 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.561827 kubelet[2879]: E0306 02:38:35.561011 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.568085 kubelet[2879]: E0306 02:38:35.566082 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.568085 kubelet[2879]: W0306 02:38:35.566159 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.568085 kubelet[2879]: E0306 02:38:35.566175 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.577457 kubelet[2879]: E0306 02:38:35.576471 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.577457 kubelet[2879]: W0306 02:38:35.576508 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.577457 kubelet[2879]: E0306 02:38:35.576543 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.585453 kubelet[2879]: E0306 02:38:35.584574 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.585453 kubelet[2879]: W0306 02:38:35.584752 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.585453 kubelet[2879]: E0306 02:38:35.584785 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.593539 kubelet[2879]: E0306 02:38:35.591016 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.593539 kubelet[2879]: W0306 02:38:35.591127 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.593539 kubelet[2879]: E0306 02:38:35.591168 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.594756 kubelet[2879]: E0306 02:38:35.594460 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.594756 kubelet[2879]: W0306 02:38:35.594543 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.594756 kubelet[2879]: E0306 02:38:35.594566 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.596586 kubelet[2879]: E0306 02:38:35.596239 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.597968 kubelet[2879]: W0306 02:38:35.597099 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.599203 kubelet[2879]: E0306 02:38:35.598157 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.639235 kubelet[2879]: E0306 02:38:35.635785 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.639235 kubelet[2879]: W0306 02:38:35.635813 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.639235 kubelet[2879]: E0306 02:38:35.635843 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.642570 kubelet[2879]: E0306 02:38:35.642213 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.642570 kubelet[2879]: W0306 02:38:35.642240 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.642570 kubelet[2879]: E0306 02:38:35.642274 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.646833 kubelet[2879]: E0306 02:38:35.644114 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.646833 kubelet[2879]: W0306 02:38:35.644204 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.646833 kubelet[2879]: E0306 02:38:35.644225 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.666829 kubelet[2879]: E0306 02:38:35.666539 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.666829 kubelet[2879]: W0306 02:38:35.666725 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.666829 kubelet[2879]: E0306 02:38:35.666760 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.679423 kubelet[2879]: E0306 02:38:35.677829 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.679584 kubelet[2879]: W0306 02:38:35.679495 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.679584 kubelet[2879]: E0306 02:38:35.679530 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.691755 kubelet[2879]: E0306 02:38:35.691076 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.691755 kubelet[2879]: W0306 02:38:35.691104 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.691755 kubelet[2879]: E0306 02:38:35.691133 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.695915 kubelet[2879]: E0306 02:38:35.695549 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.695915 kubelet[2879]: W0306 02:38:35.695731 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.695915 kubelet[2879]: E0306 02:38:35.695757 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.697906 kubelet[2879]: E0306 02:38:35.697238 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.697906 kubelet[2879]: W0306 02:38:35.697266 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.697906 kubelet[2879]: E0306 02:38:35.697489 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.700198 kubelet[2879]: E0306 02:38:35.698071 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.700198 kubelet[2879]: W0306 02:38:35.698516 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.700198 kubelet[2879]: E0306 02:38:35.698538 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.709261 kubelet[2879]: E0306 02:38:35.708956 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.709261 kubelet[2879]: W0306 02:38:35.708979 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.709261 kubelet[2879]: E0306 02:38:35.709001 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.728886 kubelet[2879]: E0306 02:38:35.728866 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.730835 kubelet[2879]: W0306 02:38:35.730098 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.730835 kubelet[2879]: E0306 02:38:35.730126 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.733511 kubelet[2879]: E0306 02:38:35.733490 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.733703 kubelet[2879]: W0306 02:38:35.733590 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.733799 kubelet[2879]: E0306 02:38:35.733778 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.748772 kubelet[2879]: E0306 02:38:35.744755 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.748772 kubelet[2879]: W0306 02:38:35.744789 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.748772 kubelet[2879]: E0306 02:38:35.744828 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.751257 kubelet[2879]: E0306 02:38:35.750716 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.751257 kubelet[2879]: W0306 02:38:35.750743 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.751257 kubelet[2879]: E0306 02:38:35.750764 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.751257 kubelet[2879]: E0306 02:38:35.751086 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.751257 kubelet[2879]: W0306 02:38:35.751098 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.751257 kubelet[2879]: E0306 02:38:35.751112 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.752748 kubelet[2879]: E0306 02:38:35.752518 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.752748 kubelet[2879]: W0306 02:38:35.752729 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.752748 kubelet[2879]: E0306 02:38:35.752751 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.756148 kubelet[2879]: E0306 02:38:35.755773 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.756148 kubelet[2879]: W0306 02:38:35.755874 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.756148 kubelet[2879]: E0306 02:38:35.755899 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.773748 kubelet[2879]: E0306 02:38:35.773545 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:35.773748 kubelet[2879]: W0306 02:38:35.773574 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:35.773748 kubelet[2879]: E0306 02:38:35.773600 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:35.789735 containerd[1569]: time="2026-03-06T02:38:35.789576757Z" level=info msg="connecting to shim 2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98" address="unix:///run/containerd/s/7e3eca65c31364f18df3682bfc7ef3becd8da69eb5f28df37c271e2b16062033" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:38:35.954973 systemd[1]: Started cri-containerd-2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98.scope - libcontainer container 2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98. Mar 6 02:38:36.071000 containerd[1569]: time="2026-03-06T02:38:36.070771881Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-75c59bd9f9-vcbwp,Uid:005aac40-e789-4801-bb37-c5b65b298482,Namespace:calico-system,Attempt:0,} returns sandbox id \"9a8cae8d13218009569f1f0a48184558c37ee12e13a4fd8c8f0fd84a3b6ba46e\"" Mar 6 02:38:36.088219 containerd[1569]: time="2026-03-06T02:38:36.087728318Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\"" Mar 6 02:38:36.146370 containerd[1569]: time="2026-03-06T02:38:36.146148682Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-cvqkx,Uid:ed53f349-d1f4-4f37-816e-cefd154747f5,Namespace:calico-system,Attempt:0,} returns sandbox id \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\"" Mar 6 02:38:37.273921 kubelet[2879]: E0306 02:38:37.272930 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:37.324588 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4105057415.mount: Deactivated successfully. Mar 6 02:38:39.271477 kubelet[2879]: E0306 02:38:39.269646 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:41.270851 kubelet[2879]: E0306 02:38:41.270027 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:43.269105 kubelet[2879]: E0306 02:38:43.268828 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:43.766823 containerd[1569]: time="2026-03-06T02:38:43.766693868Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:43.771257 containerd[1569]: time="2026-03-06T02:38:43.771011979Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.4: active requests=0, bytes read=36107596" Mar 6 02:38:43.775828 containerd[1569]: time="2026-03-06T02:38:43.775547661Z" level=info msg="ImageCreate event name:\"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:43.783561 containerd[1569]: time="2026-03-06T02:38:43.783512593Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:43.784262 containerd[1569]: time="2026-03-06T02:38:43.784085060Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.4\" with image id \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\", size \"36107450\" in 7.696310305s" Mar 6 02:38:43.784262 containerd[1569]: time="2026-03-06T02:38:43.784145634Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\" returns image reference \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\"" Mar 6 02:38:43.792214 containerd[1569]: time="2026-03-06T02:38:43.791883287Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\"" Mar 6 02:38:43.834967 containerd[1569]: time="2026-03-06T02:38:43.834845098Z" level=info msg="CreateContainer within sandbox \"9a8cae8d13218009569f1f0a48184558c37ee12e13a4fd8c8f0fd84a3b6ba46e\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Mar 6 02:38:43.861195 containerd[1569]: time="2026-03-06T02:38:43.860872087Z" level=info msg="Container b430fef9a1d7181012bfa257aa87a647c89cc59dd687887dccb02ec1dc372ced: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:38:43.901145 containerd[1569]: time="2026-03-06T02:38:43.900930092Z" level=info msg="CreateContainer within sandbox \"9a8cae8d13218009569f1f0a48184558c37ee12e13a4fd8c8f0fd84a3b6ba46e\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"b430fef9a1d7181012bfa257aa87a647c89cc59dd687887dccb02ec1dc372ced\"" Mar 6 02:38:43.908269 containerd[1569]: time="2026-03-06T02:38:43.908201619Z" level=info msg="StartContainer for \"b430fef9a1d7181012bfa257aa87a647c89cc59dd687887dccb02ec1dc372ced\"" Mar 6 02:38:43.915036 containerd[1569]: time="2026-03-06T02:38:43.914483561Z" level=info msg="connecting to shim b430fef9a1d7181012bfa257aa87a647c89cc59dd687887dccb02ec1dc372ced" address="unix:///run/containerd/s/c80a726a51b8cbf181088b40524d32551a2147593411873bf18208aab85458fc" protocol=ttrpc version=3 Mar 6 02:38:44.054169 systemd[1]: Started cri-containerd-b430fef9a1d7181012bfa257aa87a647c89cc59dd687887dccb02ec1dc372ced.scope - libcontainer container b430fef9a1d7181012bfa257aa87a647c89cc59dd687887dccb02ec1dc372ced. Mar 6 02:38:44.362039 containerd[1569]: time="2026-03-06T02:38:44.361827230Z" level=info msg="StartContainer for \"b430fef9a1d7181012bfa257aa87a647c89cc59dd687887dccb02ec1dc372ced\" returns successfully" Mar 6 02:38:44.809226 kubelet[2879]: E0306 02:38:44.804134 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.809226 kubelet[2879]: W0306 02:38:44.804168 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.809226 kubelet[2879]: E0306 02:38:44.804196 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.812659 kubelet[2879]: E0306 02:38:44.812635 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.813183 kubelet[2879]: W0306 02:38:44.813159 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.816444 kubelet[2879]: E0306 02:38:44.813447 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.817550 kubelet[2879]: E0306 02:38:44.817527 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.817653 kubelet[2879]: W0306 02:38:44.817635 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.817844 kubelet[2879]: E0306 02:38:44.817823 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.818919 kubelet[2879]: E0306 02:38:44.818855 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.818919 kubelet[2879]: W0306 02:38:44.818875 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.818919 kubelet[2879]: E0306 02:38:44.818893 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.822952 kubelet[2879]: E0306 02:38:44.821537 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.822952 kubelet[2879]: W0306 02:38:44.821809 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.822952 kubelet[2879]: E0306 02:38:44.821833 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.824046 kubelet[2879]: E0306 02:38:44.823642 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.824248 kubelet[2879]: W0306 02:38:44.824228 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.824924 kubelet[2879]: E0306 02:38:44.824906 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.836695 kubelet[2879]: E0306 02:38:44.836649 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.836695 kubelet[2879]: W0306 02:38:44.836684 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.837444 kubelet[2879]: E0306 02:38:44.836793 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.837665 kubelet[2879]: E0306 02:38:44.837640 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.837869 kubelet[2879]: W0306 02:38:44.837841 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.837957 kubelet[2879]: E0306 02:38:44.837938 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.838671 kubelet[2879]: E0306 02:38:44.838653 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.839085 kubelet[2879]: W0306 02:38:44.838989 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.839085 kubelet[2879]: E0306 02:38:44.839019 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.842129 kubelet[2879]: E0306 02:38:44.841565 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.842129 kubelet[2879]: W0306 02:38:44.841585 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.842129 kubelet[2879]: E0306 02:38:44.841605 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.844928 kubelet[2879]: E0306 02:38:44.844572 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.844928 kubelet[2879]: W0306 02:38:44.844605 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.844928 kubelet[2879]: E0306 02:38:44.844634 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.848483 kubelet[2879]: E0306 02:38:44.847982 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.848483 kubelet[2879]: W0306 02:38:44.848000 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.848483 kubelet[2879]: E0306 02:38:44.848023 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.848861 kubelet[2879]: E0306 02:38:44.848662 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.848861 kubelet[2879]: W0306 02:38:44.848679 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.848861 kubelet[2879]: E0306 02:38:44.848697 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.851989 kubelet[2879]: E0306 02:38:44.851968 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.852543 kubelet[2879]: W0306 02:38:44.852235 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.852543 kubelet[2879]: E0306 02:38:44.852269 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.853643 kubelet[2879]: E0306 02:38:44.853525 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.853643 kubelet[2879]: W0306 02:38:44.853542 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.853643 kubelet[2879]: E0306 02:38:44.853560 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.922820 kubelet[2879]: E0306 02:38:44.922625 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.922820 kubelet[2879]: W0306 02:38:44.922660 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.922820 kubelet[2879]: E0306 02:38:44.922687 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.925493 kubelet[2879]: E0306 02:38:44.925425 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.925493 kubelet[2879]: W0306 02:38:44.925450 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.925493 kubelet[2879]: E0306 02:38:44.925472 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.930546 kubelet[2879]: E0306 02:38:44.930497 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.931889 kubelet[2879]: W0306 02:38:44.931850 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.933434 kubelet[2879]: E0306 02:38:44.932520 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.939521 kubelet[2879]: E0306 02:38:44.939500 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.941691 kubelet[2879]: W0306 02:38:44.939665 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.941691 kubelet[2879]: E0306 02:38:44.941438 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.943150 kubelet[2879]: E0306 02:38:44.943132 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.943259 kubelet[2879]: W0306 02:38:44.943241 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.945614 kubelet[2879]: E0306 02:38:44.945457 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.946990 kubelet[2879]: E0306 02:38:44.946896 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.947605 kubelet[2879]: W0306 02:38:44.947577 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.948453 kubelet[2879]: E0306 02:38:44.948197 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.952621 kubelet[2879]: E0306 02:38:44.952598 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.953950 kubelet[2879]: W0306 02:38:44.953928 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.954044 kubelet[2879]: E0306 02:38:44.954031 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.956521 kubelet[2879]: E0306 02:38:44.956506 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.956813 kubelet[2879]: W0306 02:38:44.956591 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.956813 kubelet[2879]: E0306 02:38:44.956792 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.959865 kubelet[2879]: E0306 02:38:44.959494 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.959945 kubelet[2879]: W0306 02:38:44.959930 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.961499 kubelet[2879]: E0306 02:38:44.961483 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.965819 kubelet[2879]: E0306 02:38:44.965670 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.965819 kubelet[2879]: W0306 02:38:44.965690 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.965819 kubelet[2879]: E0306 02:38:44.965796 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.971524 kubelet[2879]: E0306 02:38:44.971505 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.971617 kubelet[2879]: W0306 02:38:44.971599 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.972237 kubelet[2879]: E0306 02:38:44.972217 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.981591 kubelet[2879]: E0306 02:38:44.981542 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.981591 kubelet[2879]: W0306 02:38:44.981559 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.981591 kubelet[2879]: E0306 02:38:44.981574 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.985476 kubelet[2879]: E0306 02:38:44.984707 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.985476 kubelet[2879]: W0306 02:38:44.984937 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.985476 kubelet[2879]: E0306 02:38:44.984979 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.992432 kubelet[2879]: E0306 02:38:44.989001 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.992432 kubelet[2879]: W0306 02:38:44.989029 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.992432 kubelet[2879]: E0306 02:38:44.989058 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.992432 kubelet[2879]: E0306 02:38:44.990854 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.992432 kubelet[2879]: W0306 02:38:44.990870 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.992432 kubelet[2879]: E0306 02:38:44.990893 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.992432 kubelet[2879]: E0306 02:38:44.991819 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.992432 kubelet[2879]: W0306 02:38:44.991833 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.992432 kubelet[2879]: E0306 02:38:44.991852 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:44.992432 kubelet[2879]: E0306 02:38:44.992180 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:44.992906 kubelet[2879]: W0306 02:38:44.992191 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:44.992906 kubelet[2879]: E0306 02:38:44.992205 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.000496 kubelet[2879]: E0306 02:38:44.999813 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.000496 kubelet[2879]: W0306 02:38:45.000405 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.000496 kubelet[2879]: E0306 02:38:45.000442 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.151701 containerd[1569]: time="2026-03-06T02:38:45.150193822Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:45.163199 containerd[1569]: time="2026-03-06T02:38:45.158268079Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4: active requests=0, bytes read=4630250" Mar 6 02:38:45.172176 containerd[1569]: time="2026-03-06T02:38:45.172134363Z" level=info msg="ImageCreate event name:\"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:45.180664 containerd[1569]: time="2026-03-06T02:38:45.180100851Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:38:45.185432 containerd[1569]: time="2026-03-06T02:38:45.185225763Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" with image id \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\", size \"6186255\" in 1.393114969s" Mar 6 02:38:45.185513 containerd[1569]: time="2026-03-06T02:38:45.185445476Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" returns image reference \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\"" Mar 6 02:38:45.230533 containerd[1569]: time="2026-03-06T02:38:45.229969657Z" level=info msg="CreateContainer within sandbox \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 6 02:38:45.272037 kubelet[2879]: E0306 02:38:45.271583 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:45.274478 containerd[1569]: time="2026-03-06T02:38:45.274241089Z" level=info msg="Container d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:38:45.355124 containerd[1569]: time="2026-03-06T02:38:45.348180883Z" level=info msg="CreateContainer within sandbox \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a\"" Mar 6 02:38:45.355124 containerd[1569]: time="2026-03-06T02:38:45.351239884Z" level=info msg="StartContainer for \"d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a\"" Mar 6 02:38:45.384090 containerd[1569]: time="2026-03-06T02:38:45.379013520Z" level=info msg="connecting to shim d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a" address="unix:///run/containerd/s/7e3eca65c31364f18df3682bfc7ef3becd8da69eb5f28df37c271e2b16062033" protocol=ttrpc version=3 Mar 6 02:38:45.636045 systemd[1]: Started cri-containerd-d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a.scope - libcontainer container d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a. Mar 6 02:38:45.872250 kubelet[2879]: E0306 02:38:45.871596 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.872250 kubelet[2879]: W0306 02:38:45.871634 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.872250 kubelet[2879]: E0306 02:38:45.871666 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.873194 kubelet[2879]: E0306 02:38:45.872678 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.873194 kubelet[2879]: W0306 02:38:45.872695 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.873194 kubelet[2879]: E0306 02:38:45.872813 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.879224 kubelet[2879]: E0306 02:38:45.873616 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.879224 kubelet[2879]: W0306 02:38:45.873630 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.879224 kubelet[2879]: E0306 02:38:45.873647 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.879224 kubelet[2879]: E0306 02:38:45.875010 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.879224 kubelet[2879]: W0306 02:38:45.875023 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.879224 kubelet[2879]: E0306 02:38:45.875039 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.882665 kubelet[2879]: I0306 02:38:45.882597 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-typha-75c59bd9f9-vcbwp" podStartSLOduration=4.182523284 podStartE2EDuration="11.882584119s" podCreationTimestamp="2026-03-06 02:38:34 +0000 UTC" firstStartedPulling="2026-03-06 02:38:36.087022684 +0000 UTC m=+52.094221186" lastFinishedPulling="2026-03-06 02:38:43.78708353 +0000 UTC m=+59.794282021" observedRunningTime="2026-03-06 02:38:45.04647416 +0000 UTC m=+61.053672732" watchObservedRunningTime="2026-03-06 02:38:45.882584119 +0000 UTC m=+61.889782611" Mar 6 02:38:45.889851 kubelet[2879]: E0306 02:38:45.889538 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.889851 kubelet[2879]: W0306 02:38:45.889570 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.889851 kubelet[2879]: E0306 02:38:45.889593 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.898016 kubelet[2879]: E0306 02:38:45.897626 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.898016 kubelet[2879]: W0306 02:38:45.897653 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.898016 kubelet[2879]: E0306 02:38:45.897681 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.899076 kubelet[2879]: E0306 02:38:45.899056 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.905040 kubelet[2879]: W0306 02:38:45.901912 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.905199 kubelet[2879]: E0306 02:38:45.905181 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.907710 kubelet[2879]: E0306 02:38:45.907640 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.907710 kubelet[2879]: W0306 02:38:45.907657 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.907710 kubelet[2879]: E0306 02:38:45.907679 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.912987 kubelet[2879]: E0306 02:38:45.912679 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.912987 kubelet[2879]: W0306 02:38:45.912697 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.912987 kubelet[2879]: E0306 02:38:45.912910 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.915215 kubelet[2879]: E0306 02:38:45.915198 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.916514 kubelet[2879]: W0306 02:38:45.915462 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.916514 kubelet[2879]: E0306 02:38:45.915488 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.918618 kubelet[2879]: E0306 02:38:45.918600 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.918692 kubelet[2879]: W0306 02:38:45.918680 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.918843 kubelet[2879]: E0306 02:38:45.918830 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.919603 kubelet[2879]: E0306 02:38:45.919590 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.919660 kubelet[2879]: W0306 02:38:45.919650 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.919823 kubelet[2879]: E0306 02:38:45.919713 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.920213 kubelet[2879]: E0306 02:38:45.920148 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.920213 kubelet[2879]: W0306 02:38:45.920161 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.920213 kubelet[2879]: E0306 02:38:45.920173 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.926075 kubelet[2879]: E0306 02:38:45.925976 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.926075 kubelet[2879]: W0306 02:38:45.925993 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.926075 kubelet[2879]: E0306 02:38:45.926008 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.928879 kubelet[2879]: E0306 02:38:45.928671 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.928945 kubelet[2879]: W0306 02:38:45.928878 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.928945 kubelet[2879]: E0306 02:38:45.928920 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.960667 kubelet[2879]: E0306 02:38:45.959467 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.960667 kubelet[2879]: W0306 02:38:45.959578 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.960667 kubelet[2879]: E0306 02:38:45.959613 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.963069 kubelet[2879]: E0306 02:38:45.962463 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.963069 kubelet[2879]: W0306 02:38:45.962565 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.963069 kubelet[2879]: E0306 02:38:45.962588 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.965965 kubelet[2879]: E0306 02:38:45.965674 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.965965 kubelet[2879]: W0306 02:38:45.965864 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.965965 kubelet[2879]: E0306 02:38:45.965885 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.971634 kubelet[2879]: E0306 02:38:45.971474 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.971634 kubelet[2879]: W0306 02:38:45.971504 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.971634 kubelet[2879]: E0306 02:38:45.971528 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.975613 kubelet[2879]: E0306 02:38:45.975140 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.975613 kubelet[2879]: W0306 02:38:45.975252 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.975613 kubelet[2879]: E0306 02:38:45.975444 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.978508 kubelet[2879]: E0306 02:38:45.978142 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.978508 kubelet[2879]: W0306 02:38:45.978231 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.978508 kubelet[2879]: E0306 02:38:45.978257 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.981901 kubelet[2879]: E0306 02:38:45.981101 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.981901 kubelet[2879]: W0306 02:38:45.981127 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.981901 kubelet[2879]: E0306 02:38:45.981169 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.986895 kubelet[2879]: E0306 02:38:45.986874 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.986998 kubelet[2879]: W0306 02:38:45.986980 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.987074 kubelet[2879]: E0306 02:38:45.987060 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.988603 kubelet[2879]: E0306 02:38:45.988578 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.988890 kubelet[2879]: W0306 02:38:45.988868 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.989029 kubelet[2879]: E0306 02:38:45.988975 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.990904 kubelet[2879]: E0306 02:38:45.990886 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.991002 kubelet[2879]: W0306 02:38:45.990983 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.991082 kubelet[2879]: E0306 02:38:45.991066 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.996175 kubelet[2879]: E0306 02:38:45.995918 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.996175 kubelet[2879]: W0306 02:38:45.996020 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.996175 kubelet[2879]: E0306 02:38:45.996061 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.997571 kubelet[2879]: E0306 02:38:45.997187 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.997571 kubelet[2879]: W0306 02:38:45.997420 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.997571 kubelet[2879]: E0306 02:38:45.997453 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:45.999555 kubelet[2879]: E0306 02:38:45.999445 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:45.999555 kubelet[2879]: W0306 02:38:45.999537 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:45.999658 kubelet[2879]: E0306 02:38:45.999564 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:46.001888 kubelet[2879]: E0306 02:38:46.001546 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:46.001888 kubelet[2879]: W0306 02:38:46.001573 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:46.001888 kubelet[2879]: E0306 02:38:46.001588 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:46.007156 kubelet[2879]: E0306 02:38:46.006058 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:46.007156 kubelet[2879]: W0306 02:38:46.006073 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:46.007156 kubelet[2879]: E0306 02:38:46.006088 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:46.010114 kubelet[2879]: E0306 02:38:46.009859 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:46.010114 kubelet[2879]: W0306 02:38:46.010029 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:46.010114 kubelet[2879]: E0306 02:38:46.010050 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:46.012520 kubelet[2879]: E0306 02:38:46.012135 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:46.012520 kubelet[2879]: W0306 02:38:46.012229 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:46.012520 kubelet[2879]: E0306 02:38:46.012246 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:46.014062 kubelet[2879]: E0306 02:38:46.013940 2879 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 6 02:38:46.014062 kubelet[2879]: W0306 02:38:46.013961 2879 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 6 02:38:46.014062 kubelet[2879]: E0306 02:38:46.013978 2879 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 6 02:38:46.172518 systemd[1]: cri-containerd-d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a.scope: Deactivated successfully. Mar 6 02:38:46.182239 containerd[1569]: time="2026-03-06T02:38:46.182097388Z" level=info msg="StartContainer for \"d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a\" returns successfully" Mar 6 02:38:46.207474 containerd[1569]: time="2026-03-06T02:38:46.207184259Z" level=info msg="received container exit event container_id:\"d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a\" id:\"d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a\" pid:3578 exited_at:{seconds:1772764726 nanos:193507760}" Mar 6 02:38:46.339585 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d0848f66d27d02228705eefdfb8ce58d0d655045a1fc004798f20709bf69bb4a-rootfs.mount: Deactivated successfully. Mar 6 02:38:46.826527 containerd[1569]: time="2026-03-06T02:38:46.825035775Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\"" Mar 6 02:38:47.270972 kubelet[2879]: E0306 02:38:47.270231 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:49.270528 kubelet[2879]: E0306 02:38:49.269157 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:51.270520 kubelet[2879]: E0306 02:38:51.270274 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:53.272563 kubelet[2879]: E0306 02:38:53.271063 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:55.269273 kubelet[2879]: E0306 02:38:55.268626 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:57.271175 kubelet[2879]: E0306 02:38:57.270759 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:38:59.273551 kubelet[2879]: E0306 02:38:59.269688 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:01.270425 kubelet[2879]: E0306 02:39:01.270008 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:03.288100 kubelet[2879]: E0306 02:39:03.287275 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:05.270703 kubelet[2879]: E0306 02:39:05.269656 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:07.270273 kubelet[2879]: E0306 02:39:07.269788 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:09.272906 kubelet[2879]: E0306 02:39:09.272092 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:11.271264 kubelet[2879]: E0306 02:39:11.269837 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:13.275784 kubelet[2879]: E0306 02:39:13.269909 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:15.648722 kubelet[2879]: E0306 02:39:15.647484 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:17.270097 kubelet[2879]: E0306 02:39:17.269927 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:19.269218 kubelet[2879]: E0306 02:39:19.268701 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:20.161799 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3817314823.mount: Deactivated successfully. Mar 6 02:39:20.282448 containerd[1569]: time="2026-03-06T02:39:20.281725118Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:20.285803 containerd[1569]: time="2026-03-06T02:39:20.284503653Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.4: active requests=0, bytes read=159838564" Mar 6 02:39:20.291004 containerd[1569]: time="2026-03-06T02:39:20.290613187Z" level=info msg="ImageCreate event name:\"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:20.374155 containerd[1569]: time="2026-03-06T02:39:20.372674644Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:20.380787 containerd[1569]: time="2026-03-06T02:39:20.377543377Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.4\" with image id \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\", size \"159838426\" in 33.552447528s" Mar 6 02:39:20.380787 containerd[1569]: time="2026-03-06T02:39:20.377711011Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\" returns image reference \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\"" Mar 6 02:39:20.426549 containerd[1569]: time="2026-03-06T02:39:20.423904698Z" level=info msg="CreateContainer within sandbox \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Mar 6 02:39:20.492448 containerd[1569]: time="2026-03-06T02:39:20.492174818Z" level=info msg="Container ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:39:20.510818 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2580142284.mount: Deactivated successfully. Mar 6 02:39:20.587586 containerd[1569]: time="2026-03-06T02:39:20.585839351Z" level=info msg="CreateContainer within sandbox \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6\"" Mar 6 02:39:20.589217 containerd[1569]: time="2026-03-06T02:39:20.588518531Z" level=info msg="StartContainer for \"ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6\"" Mar 6 02:39:20.593391 containerd[1569]: time="2026-03-06T02:39:20.592944752Z" level=info msg="connecting to shim ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6" address="unix:///run/containerd/s/7e3eca65c31364f18df3682bfc7ef3becd8da69eb5f28df37c271e2b16062033" protocol=ttrpc version=3 Mar 6 02:39:20.741442 systemd[1]: Started cri-containerd-ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6.scope - libcontainer container ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6. Mar 6 02:39:21.227472 containerd[1569]: time="2026-03-06T02:39:21.226822365Z" level=info msg="StartContainer for \"ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6\" returns successfully" Mar 6 02:39:21.269716 kubelet[2879]: E0306 02:39:21.269545 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:21.572223 systemd[1]: cri-containerd-ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6.scope: Deactivated successfully. Mar 6 02:39:21.628431 containerd[1569]: time="2026-03-06T02:39:21.625195191Z" level=info msg="received container exit event container_id:\"ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6\" id:\"ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6\" pid:3677 exited_at:{seconds:1772764761 nanos:604685445}" Mar 6 02:39:21.850469 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ecd1a1b505b91b6bd880d92b2e8d20b6fe4e5e86ac66b32797a865aecafe1ef6-rootfs.mount: Deactivated successfully. Mar 6 02:39:22.946772 containerd[1569]: time="2026-03-06T02:39:22.946663100Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\"" Mar 6 02:39:23.275977 kubelet[2879]: E0306 02:39:23.271654 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:25.271722 kubelet[2879]: E0306 02:39:25.269142 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:27.278271 kubelet[2879]: E0306 02:39:27.276559 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:29.276175 kubelet[2879]: E0306 02:39:29.269638 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:31.277681 kubelet[2879]: E0306 02:39:31.276852 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:33.268964 kubelet[2879]: E0306 02:39:33.268455 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:35.271211 kubelet[2879]: E0306 02:39:35.270909 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:36.698507 containerd[1569]: time="2026-03-06T02:39:36.698068996Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:36.701967 containerd[1569]: time="2026-03-06T02:39:36.701014872Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.4: active requests=0, bytes read=70611671" Mar 6 02:39:36.761735 containerd[1569]: time="2026-03-06T02:39:36.761027789Z" level=info msg="ImageCreate event name:\"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:36.768504 containerd[1569]: time="2026-03-06T02:39:36.768028096Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:36.769432 containerd[1569]: time="2026-03-06T02:39:36.768858045Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.4\" with image id \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\", size \"72167716\" in 13.817689205s" Mar 6 02:39:36.769432 containerd[1569]: time="2026-03-06T02:39:36.768983190Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\" returns image reference \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\"" Mar 6 02:39:36.822464 containerd[1569]: time="2026-03-06T02:39:36.821788270Z" level=info msg="CreateContainer within sandbox \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 6 02:39:36.863731 containerd[1569]: time="2026-03-06T02:39:36.863676376Z" level=info msg="Container 06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:39:36.933394 containerd[1569]: time="2026-03-06T02:39:36.933028583Z" level=info msg="CreateContainer within sandbox \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86\"" Mar 6 02:39:36.936434 containerd[1569]: time="2026-03-06T02:39:36.935575712Z" level=info msg="StartContainer for \"06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86\"" Mar 6 02:39:36.939527 containerd[1569]: time="2026-03-06T02:39:36.938981117Z" level=info msg="connecting to shim 06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86" address="unix:///run/containerd/s/7e3eca65c31364f18df3682bfc7ef3becd8da69eb5f28df37c271e2b16062033" protocol=ttrpc version=3 Mar 6 02:39:37.022972 systemd[1]: Started cri-containerd-06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86.scope - libcontainer container 06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86. Mar 6 02:39:37.269133 kubelet[2879]: E0306 02:39:37.269081 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:37.440957 containerd[1569]: time="2026-03-06T02:39:37.439896133Z" level=info msg="StartContainer for \"06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86\" returns successfully" Mar 6 02:39:39.271835 kubelet[2879]: E0306 02:39:39.269738 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:39.974699 systemd[1]: cri-containerd-06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86.scope: Deactivated successfully. Mar 6 02:39:39.976056 systemd[1]: cri-containerd-06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86.scope: Consumed 2.097s CPU time, 181M memory peak, 4.2M read from disk, 177M written to disk. Mar 6 02:39:39.979887 containerd[1569]: time="2026-03-06T02:39:39.979740232Z" level=info msg="received container exit event container_id:\"06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86\" id:\"06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86\" pid:3739 exited_at:{seconds:1772764779 nanos:977048016}" Mar 6 02:39:40.029716 kubelet[2879]: I0306 02:39:40.029028 2879 kubelet_node_status.go:427] "Fast updating node status as it just became ready" Mar 6 02:39:40.188863 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-06c44a02812e92dc73ec80c3f2e29b9c95dce7da8c2fcb5319e3ab80139feb86-rootfs.mount: Deactivated successfully. Mar 6 02:39:40.551958 systemd[1]: Created slice kubepods-burstable-pod48337397_07a0_44d5_b732_338087eee7ba.slice - libcontainer container kubepods-burstable-pod48337397_07a0_44d5_b732_338087eee7ba.slice. Mar 6 02:39:40.600983 systemd[1]: Created slice kubepods-besteffort-podbe2e66c8_0f54_46c3_af38_09952dbbac49.slice - libcontainer container kubepods-besteffort-podbe2e66c8_0f54_46c3_af38_09952dbbac49.slice. Mar 6 02:39:40.638445 systemd[1]: Created slice kubepods-besteffort-pod64f79500_120a_4333_adde_31efeaa36e77.slice - libcontainer container kubepods-besteffort-pod64f79500_120a_4333_adde_31efeaa36e77.slice. Mar 6 02:39:40.666249 kubelet[2879]: I0306 02:39:40.666037 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64f79500-120a-4333-adde-31efeaa36e77-config\") pod \"goldmane-9f7667bb8-bz8j9\" (UID: \"64f79500-120a-4333-adde-31efeaa36e77\") " pod="calico-system/goldmane-9f7667bb8-bz8j9" Mar 6 02:39:40.666929 kubelet[2879]: I0306 02:39:40.666271 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-ca-bundle\") pod \"whisker-794b7cb656-6plgr\" (UID: \"ee2caea9-f1f8-4884-8243-566cb757b833\") " pod="calico-system/whisker-794b7cb656-6plgr" Mar 6 02:39:40.666929 kubelet[2879]: I0306 02:39:40.666477 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbc9t\" (UniqueName: \"kubernetes.io/projected/ee80ff4d-1458-4649-bc23-11b25825a7c0-kube-api-access-kbc9t\") pod \"coredns-7d764666f9-r6hk4\" (UID: \"ee80ff4d-1458-4649-bc23-11b25825a7c0\") " pod="kube-system/coredns-7d764666f9-r6hk4" Mar 6 02:39:40.666929 kubelet[2879]: I0306 02:39:40.666504 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64f79500-120a-4333-adde-31efeaa36e77-goldmane-ca-bundle\") pod \"goldmane-9f7667bb8-bz8j9\" (UID: \"64f79500-120a-4333-adde-31efeaa36e77\") " pod="calico-system/goldmane-9f7667bb8-bz8j9" Mar 6 02:39:40.666929 kubelet[2879]: I0306 02:39:40.666523 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f-calico-apiserver-certs\") pod \"calico-apiserver-d54f5f4cf-t2lk8\" (UID: \"5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f\") " pod="calico-system/calico-apiserver-d54f5f4cf-t2lk8" Mar 6 02:39:40.666929 kubelet[2879]: I0306 02:39:40.666542 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwps\" (UniqueName: \"kubernetes.io/projected/5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f-kube-api-access-djwps\") pod \"calico-apiserver-d54f5f4cf-t2lk8\" (UID: \"5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f\") " pod="calico-system/calico-apiserver-d54f5f4cf-t2lk8" Mar 6 02:39:40.667120 kubelet[2879]: I0306 02:39:40.666566 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-nginx-config\") pod \"whisker-794b7cb656-6plgr\" (UID: \"ee2caea9-f1f8-4884-8243-566cb757b833\") " pod="calico-system/whisker-794b7cb656-6plgr" Mar 6 02:39:40.667120 kubelet[2879]: I0306 02:39:40.666598 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee80ff4d-1458-4649-bc23-11b25825a7c0-config-volume\") pod \"coredns-7d764666f9-r6hk4\" (UID: \"ee80ff4d-1458-4649-bc23-11b25825a7c0\") " pod="kube-system/coredns-7d764666f9-r6hk4" Mar 6 02:39:40.667120 kubelet[2879]: I0306 02:39:40.666618 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g2vf\" (UniqueName: \"kubernetes.io/projected/48337397-07a0-44d5-b732-338087eee7ba-kube-api-access-7g2vf\") pod \"coredns-7d764666f9-rghxd\" (UID: \"48337397-07a0-44d5-b732-338087eee7ba\") " pod="kube-system/coredns-7d764666f9-rghxd" Mar 6 02:39:40.667120 kubelet[2879]: I0306 02:39:40.666641 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mhjv\" (UniqueName: \"kubernetes.io/projected/64f79500-120a-4333-adde-31efeaa36e77-kube-api-access-5mhjv\") pod \"goldmane-9f7667bb8-bz8j9\" (UID: \"64f79500-120a-4333-adde-31efeaa36e77\") " pod="calico-system/goldmane-9f7667bb8-bz8j9" Mar 6 02:39:40.667120 kubelet[2879]: I0306 02:39:40.666661 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/be2e66c8-0f54-46c3-af38-09952dbbac49-calico-apiserver-certs\") pod \"calico-apiserver-d54f5f4cf-vrw2m\" (UID: \"be2e66c8-0f54-46c3-af38-09952dbbac49\") " pod="calico-system/calico-apiserver-d54f5f4cf-vrw2m" Mar 6 02:39:40.675523 kubelet[2879]: I0306 02:39:40.666682 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mfxj\" (UniqueName: \"kubernetes.io/projected/be2e66c8-0f54-46c3-af38-09952dbbac49-kube-api-access-4mfxj\") pod \"calico-apiserver-d54f5f4cf-vrw2m\" (UID: \"be2e66c8-0f54-46c3-af38-09952dbbac49\") " pod="calico-system/calico-apiserver-d54f5f4cf-vrw2m" Mar 6 02:39:40.675523 kubelet[2879]: I0306 02:39:40.666706 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/64f79500-120a-4333-adde-31efeaa36e77-goldmane-key-pair\") pod \"goldmane-9f7667bb8-bz8j9\" (UID: \"64f79500-120a-4333-adde-31efeaa36e77\") " pod="calico-system/goldmane-9f7667bb8-bz8j9" Mar 6 02:39:40.675523 kubelet[2879]: I0306 02:39:40.666732 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-backend-key-pair\") pod \"whisker-794b7cb656-6plgr\" (UID: \"ee2caea9-f1f8-4884-8243-566cb757b833\") " pod="calico-system/whisker-794b7cb656-6plgr" Mar 6 02:39:40.675523 kubelet[2879]: I0306 02:39:40.666751 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvhjv\" (UniqueName: \"kubernetes.io/projected/ee2caea9-f1f8-4884-8243-566cb757b833-kube-api-access-gvhjv\") pod \"whisker-794b7cb656-6plgr\" (UID: \"ee2caea9-f1f8-4884-8243-566cb757b833\") " pod="calico-system/whisker-794b7cb656-6plgr" Mar 6 02:39:40.675523 kubelet[2879]: I0306 02:39:40.666770 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48337397-07a0-44d5-b732-338087eee7ba-config-volume\") pod \"coredns-7d764666f9-rghxd\" (UID: \"48337397-07a0-44d5-b732-338087eee7ba\") " pod="kube-system/coredns-7d764666f9-rghxd" Mar 6 02:39:40.675694 systemd[1]: Created slice kubepods-besteffort-podee2caea9_f1f8_4884_8243_566cb757b833.slice - libcontainer container kubepods-besteffort-podee2caea9_f1f8_4884_8243_566cb757b833.slice. Mar 6 02:39:40.696042 systemd[1]: Created slice kubepods-besteffort-pod5b1f7233_0ca1_4dd4_8a21_8dd7eac72e6f.slice - libcontainer container kubepods-besteffort-pod5b1f7233_0ca1_4dd4_8a21_8dd7eac72e6f.slice. Mar 6 02:39:40.728481 systemd[1]: Created slice kubepods-burstable-podee80ff4d_1458_4649_bc23_11b25825a7c0.slice - libcontainer container kubepods-burstable-podee80ff4d_1458_4649_bc23_11b25825a7c0.slice. Mar 6 02:39:40.746461 systemd[1]: Created slice kubepods-besteffort-pod2776157f_9f5c_4064_8095_991904ea8002.slice - libcontainer container kubepods-besteffort-pod2776157f_9f5c_4064_8095_991904ea8002.slice. Mar 6 02:39:40.768549 kubelet[2879]: I0306 02:39:40.767133 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnxrk\" (UniqueName: \"kubernetes.io/projected/2776157f-9f5c-4064-8095-991904ea8002-kube-api-access-tnxrk\") pod \"calico-kube-controllers-5cf8bb7c94-m8khs\" (UID: \"2776157f-9f5c-4064-8095-991904ea8002\") " pod="calico-system/calico-kube-controllers-5cf8bb7c94-m8khs" Mar 6 02:39:40.770971 kubelet[2879]: I0306 02:39:40.770936 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2776157f-9f5c-4064-8095-991904ea8002-tigera-ca-bundle\") pod \"calico-kube-controllers-5cf8bb7c94-m8khs\" (UID: \"2776157f-9f5c-4064-8095-991904ea8002\") " pod="calico-system/calico-kube-controllers-5cf8bb7c94-m8khs" Mar 6 02:39:40.940073 containerd[1569]: time="2026-03-06T02:39:40.939692635Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-d54f5f4cf-vrw2m,Uid:be2e66c8-0f54-46c3-af38-09952dbbac49,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:40.984892 containerd[1569]: time="2026-03-06T02:39:40.984743454Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-9f7667bb8-bz8j9,Uid:64f79500-120a-4333-adde-31efeaa36e77,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:40.990670 containerd[1569]: time="2026-03-06T02:39:40.989466998Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-794b7cb656-6plgr,Uid:ee2caea9-f1f8-4884-8243-566cb757b833,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:41.036449 containerd[1569]: time="2026-03-06T02:39:41.035609498Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-d54f5f4cf-t2lk8,Uid:5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:41.046044 containerd[1569]: time="2026-03-06T02:39:41.046000636Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-r6hk4,Uid:ee80ff4d-1458-4649-bc23-11b25825a7c0,Namespace:kube-system,Attempt:0,}" Mar 6 02:39:41.062659 containerd[1569]: time="2026-03-06T02:39:41.061687502Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5cf8bb7c94-m8khs,Uid:2776157f-9f5c-4064-8095-991904ea8002,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:41.262622 containerd[1569]: time="2026-03-06T02:39:41.259822216Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-rghxd,Uid:48337397-07a0-44d5-b732-338087eee7ba,Namespace:kube-system,Attempt:0,}" Mar 6 02:39:41.295259 systemd[1]: Created slice kubepods-besteffort-pod74be49b3_6c58_4ae9_86a2_c8139f48d9c9.slice - libcontainer container kubepods-besteffort-pod74be49b3_6c58_4ae9_86a2_c8139f48d9c9.slice. Mar 6 02:39:41.335634 containerd[1569]: time="2026-03-06T02:39:41.335489880Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-gqfqd,Uid:74be49b3-6c58-4ae9-86a2-c8139f48d9c9,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:41.481862 containerd[1569]: time="2026-03-06T02:39:41.481813545Z" level=info msg="CreateContainer within sandbox \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 6 02:39:41.593959 containerd[1569]: time="2026-03-06T02:39:41.593907856Z" level=info msg="Container b3a703a6ccaa7a44ddca005f1c20cecd9636e807b4d8a348944b76b31df5487b: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:39:41.741710 containerd[1569]: time="2026-03-06T02:39:41.740969819Z" level=info msg="CreateContainer within sandbox \"2c30aca3d49f06efb8826422af9967fad1e4983686a330eb9611ddd713335b98\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"b3a703a6ccaa7a44ddca005f1c20cecd9636e807b4d8a348944b76b31df5487b\"" Mar 6 02:39:41.763128 containerd[1569]: time="2026-03-06T02:39:41.763080731Z" level=info msg="StartContainer for \"b3a703a6ccaa7a44ddca005f1c20cecd9636e807b4d8a348944b76b31df5487b\"" Mar 6 02:39:41.766605 containerd[1569]: time="2026-03-06T02:39:41.766572588Z" level=info msg="connecting to shim b3a703a6ccaa7a44ddca005f1c20cecd9636e807b4d8a348944b76b31df5487b" address="unix:///run/containerd/s/7e3eca65c31364f18df3682bfc7ef3becd8da69eb5f28df37c271e2b16062033" protocol=ttrpc version=3 Mar 6 02:39:42.036048 systemd[1]: Started cri-containerd-b3a703a6ccaa7a44ddca005f1c20cecd9636e807b4d8a348944b76b31df5487b.scope - libcontainer container b3a703a6ccaa7a44ddca005f1c20cecd9636e807b4d8a348944b76b31df5487b. Mar 6 02:39:42.380496 containerd[1569]: time="2026-03-06T02:39:42.378898690Z" level=error msg="Failed to destroy network for sandbox \"a5c47e0dbc6faac54dc3ef947b78139a749e8493ce1beb670419096c8733136d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.395558 systemd[1]: run-netns-cni\x2df4414cd1\x2dc91f\x2d73a1\x2d6401\x2d470fbe341367.mount: Deactivated successfully. Mar 6 02:39:42.432075 containerd[1569]: time="2026-03-06T02:39:42.431629662Z" level=error msg="Failed to destroy network for sandbox \"e3d4684da0bfc10c9f276821ece784b4da13459a9a11c0de12de5f83b95c4191\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.436642 systemd[1]: run-netns-cni\x2d5b4e2646\x2de102\x2d849e\x2d7bae\x2dfcdce06f7d04.mount: Deactivated successfully. Mar 6 02:39:42.448867 containerd[1569]: time="2026-03-06T02:39:42.448717272Z" level=error msg="Failed to destroy network for sandbox \"2a1125d2f4d4857c791366d7adecd041c835ae57c11255e2ce92d3a199b38c0b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.451799 containerd[1569]: time="2026-03-06T02:39:42.451736712Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-794b7cb656-6plgr,Uid:ee2caea9-f1f8-4884-8243-566cb757b833,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5c47e0dbc6faac54dc3ef947b78139a749e8493ce1beb670419096c8733136d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.453037 systemd[1]: run-netns-cni\x2d5cb57a9b\x2dbe00\x2d48be\x2d43b8\x2dba7b9fdacfb7.mount: Deactivated successfully. Mar 6 02:39:42.457460 containerd[1569]: time="2026-03-06T02:39:42.456992991Z" level=error msg="Failed to destroy network for sandbox \"88ba41bad61d6a587c9468068dfbd1f3e97168c29d0089c75f0a72b13eac260c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.462865 kubelet[2879]: E0306 02:39:42.462645 2879 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5c47e0dbc6faac54dc3ef947b78139a749e8493ce1beb670419096c8733136d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.462865 kubelet[2879]: E0306 02:39:42.462804 2879 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5c47e0dbc6faac54dc3ef947b78139a749e8493ce1beb670419096c8733136d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-794b7cb656-6plgr" Mar 6 02:39:42.462865 kubelet[2879]: E0306 02:39:42.462830 2879 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5c47e0dbc6faac54dc3ef947b78139a749e8493ce1beb670419096c8733136d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-794b7cb656-6plgr" Mar 6 02:39:42.463693 kubelet[2879]: E0306 02:39:42.462890 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-794b7cb656-6plgr_calico-system(ee2caea9-f1f8-4884-8243-566cb757b833)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-794b7cb656-6plgr_calico-system(ee2caea9-f1f8-4884-8243-566cb757b833)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a5c47e0dbc6faac54dc3ef947b78139a749e8493ce1beb670419096c8733136d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-794b7cb656-6plgr" podUID="ee2caea9-f1f8-4884-8243-566cb757b833" Mar 6 02:39:42.466533 systemd[1]: run-netns-cni\x2d32a1b09c\x2d6527\x2ddb55\x2d89fd\x2dc71a59837573.mount: Deactivated successfully. Mar 6 02:39:42.474844 containerd[1569]: time="2026-03-06T02:39:42.474776793Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-rghxd,Uid:48337397-07a0-44d5-b732-338087eee7ba,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3d4684da0bfc10c9f276821ece784b4da13459a9a11c0de12de5f83b95c4191\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.481746 containerd[1569]: time="2026-03-06T02:39:42.480613089Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-r6hk4,Uid:ee80ff4d-1458-4649-bc23-11b25825a7c0,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a1125d2f4d4857c791366d7adecd041c835ae57c11255e2ce92d3a199b38c0b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.486020 containerd[1569]: time="2026-03-06T02:39:42.485618339Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-d54f5f4cf-vrw2m,Uid:be2e66c8-0f54-46c3-af38-09952dbbac49,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"88ba41bad61d6a587c9468068dfbd1f3e97168c29d0089c75f0a72b13eac260c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.487801 kubelet[2879]: E0306 02:39:42.486795 2879 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88ba41bad61d6a587c9468068dfbd1f3e97168c29d0089c75f0a72b13eac260c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.487801 kubelet[2879]: E0306 02:39:42.486958 2879 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88ba41bad61d6a587c9468068dfbd1f3e97168c29d0089c75f0a72b13eac260c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-d54f5f4cf-vrw2m" Mar 6 02:39:42.487801 kubelet[2879]: E0306 02:39:42.486988 2879 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88ba41bad61d6a587c9468068dfbd1f3e97168c29d0089c75f0a72b13eac260c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-d54f5f4cf-vrw2m" Mar 6 02:39:42.487992 kubelet[2879]: E0306 02:39:42.487053 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-d54f5f4cf-vrw2m_calico-system(be2e66c8-0f54-46c3-af38-09952dbbac49)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-d54f5f4cf-vrw2m_calico-system(be2e66c8-0f54-46c3-af38-09952dbbac49)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"88ba41bad61d6a587c9468068dfbd1f3e97168c29d0089c75f0a72b13eac260c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-d54f5f4cf-vrw2m" podUID="be2e66c8-0f54-46c3-af38-09952dbbac49" Mar 6 02:39:42.487992 kubelet[2879]: E0306 02:39:42.487136 2879 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3d4684da0bfc10c9f276821ece784b4da13459a9a11c0de12de5f83b95c4191\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.487992 kubelet[2879]: E0306 02:39:42.487262 2879 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3d4684da0bfc10c9f276821ece784b4da13459a9a11c0de12de5f83b95c4191\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-rghxd" Mar 6 02:39:42.489086 kubelet[2879]: E0306 02:39:42.487758 2879 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a1125d2f4d4857c791366d7adecd041c835ae57c11255e2ce92d3a199b38c0b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.489086 kubelet[2879]: E0306 02:39:42.489052 2879 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a1125d2f4d4857c791366d7adecd041c835ae57c11255e2ce92d3a199b38c0b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-r6hk4" Mar 6 02:39:42.489086 kubelet[2879]: E0306 02:39:42.489078 2879 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a1125d2f4d4857c791366d7adecd041c835ae57c11255e2ce92d3a199b38c0b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-r6hk4" Mar 6 02:39:42.489526 kubelet[2879]: E0306 02:39:42.489137 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7d764666f9-r6hk4_kube-system(ee80ff4d-1458-4649-bc23-11b25825a7c0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7d764666f9-r6hk4_kube-system(ee80ff4d-1458-4649-bc23-11b25825a7c0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2a1125d2f4d4857c791366d7adecd041c835ae57c11255e2ce92d3a199b38c0b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7d764666f9-r6hk4" podUID="ee80ff4d-1458-4649-bc23-11b25825a7c0" Mar 6 02:39:42.490680 kubelet[2879]: E0306 02:39:42.489990 2879 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3d4684da0bfc10c9f276821ece784b4da13459a9a11c0de12de5f83b95c4191\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-rghxd" Mar 6 02:39:42.493045 kubelet[2879]: E0306 02:39:42.492912 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7d764666f9-rghxd_kube-system(48337397-07a0-44d5-b732-338087eee7ba)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7d764666f9-rghxd_kube-system(48337397-07a0-44d5-b732-338087eee7ba)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e3d4684da0bfc10c9f276821ece784b4da13459a9a11c0de12de5f83b95c4191\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7d764666f9-rghxd" podUID="48337397-07a0-44d5-b732-338087eee7ba" Mar 6 02:39:42.494724 containerd[1569]: time="2026-03-06T02:39:42.493982065Z" level=error msg="Failed to destroy network for sandbox \"3ce681ae066fca4d3eea7c602efde0c1002d6973e560cc4d635baf7c5cb95efb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.527544 containerd[1569]: time="2026-03-06T02:39:42.527470296Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5cf8bb7c94-m8khs,Uid:2776157f-9f5c-4064-8095-991904ea8002,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3ce681ae066fca4d3eea7c602efde0c1002d6973e560cc4d635baf7c5cb95efb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.534576 kubelet[2879]: E0306 02:39:42.533916 2879 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3ce681ae066fca4d3eea7c602efde0c1002d6973e560cc4d635baf7c5cb95efb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.534576 kubelet[2879]: E0306 02:39:42.534073 2879 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3ce681ae066fca4d3eea7c602efde0c1002d6973e560cc4d635baf7c5cb95efb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5cf8bb7c94-m8khs" Mar 6 02:39:42.534776 kubelet[2879]: E0306 02:39:42.534102 2879 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3ce681ae066fca4d3eea7c602efde0c1002d6973e560cc4d635baf7c5cb95efb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5cf8bb7c94-m8khs" Mar 6 02:39:42.534837 kubelet[2879]: E0306 02:39:42.534769 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5cf8bb7c94-m8khs_calico-system(2776157f-9f5c-4064-8095-991904ea8002)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5cf8bb7c94-m8khs_calico-system(2776157f-9f5c-4064-8095-991904ea8002)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3ce681ae066fca4d3eea7c602efde0c1002d6973e560cc4d635baf7c5cb95efb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5cf8bb7c94-m8khs" podUID="2776157f-9f5c-4064-8095-991904ea8002" Mar 6 02:39:42.556458 containerd[1569]: time="2026-03-06T02:39:42.555662702Z" level=error msg="Failed to destroy network for sandbox \"137405d1f2bf1b106540b03349620738e3d1375c870a819aa01990643217bb71\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.563143 containerd[1569]: time="2026-03-06T02:39:42.562931709Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-d54f5f4cf-t2lk8,Uid:5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"137405d1f2bf1b106540b03349620738e3d1375c870a819aa01990643217bb71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.565124 kubelet[2879]: E0306 02:39:42.564146 2879 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"137405d1f2bf1b106540b03349620738e3d1375c870a819aa01990643217bb71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.565124 kubelet[2879]: E0306 02:39:42.564545 2879 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"137405d1f2bf1b106540b03349620738e3d1375c870a819aa01990643217bb71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-d54f5f4cf-t2lk8" Mar 6 02:39:42.565124 kubelet[2879]: E0306 02:39:42.564573 2879 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"137405d1f2bf1b106540b03349620738e3d1375c870a819aa01990643217bb71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-d54f5f4cf-t2lk8" Mar 6 02:39:42.565766 kubelet[2879]: E0306 02:39:42.564638 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-d54f5f4cf-t2lk8_calico-system(5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-d54f5f4cf-t2lk8_calico-system(5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"137405d1f2bf1b106540b03349620738e3d1375c870a819aa01990643217bb71\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-d54f5f4cf-t2lk8" podUID="5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f" Mar 6 02:39:42.575465 containerd[1569]: time="2026-03-06T02:39:42.573850517Z" level=error msg="Failed to destroy network for sandbox \"56c6b6ac2ca7174695af317f44ecdf75f71cc877eafa526461a583b8c56a0587\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.596636 containerd[1569]: time="2026-03-06T02:39:42.594746544Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-9f7667bb8-bz8j9,Uid:64f79500-120a-4333-adde-31efeaa36e77,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"56c6b6ac2ca7174695af317f44ecdf75f71cc877eafa526461a583b8c56a0587\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.596982 kubelet[2879]: E0306 02:39:42.595985 2879 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"56c6b6ac2ca7174695af317f44ecdf75f71cc877eafa526461a583b8c56a0587\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.596982 kubelet[2879]: E0306 02:39:42.596055 2879 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"56c6b6ac2ca7174695af317f44ecdf75f71cc877eafa526461a583b8c56a0587\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-9f7667bb8-bz8j9" Mar 6 02:39:42.596982 kubelet[2879]: E0306 02:39:42.596083 2879 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"56c6b6ac2ca7174695af317f44ecdf75f71cc877eafa526461a583b8c56a0587\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-9f7667bb8-bz8j9" Mar 6 02:39:42.597863 kubelet[2879]: E0306 02:39:42.596141 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-9f7667bb8-bz8j9_calico-system(64f79500-120a-4333-adde-31efeaa36e77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-9f7667bb8-bz8j9_calico-system(64f79500-120a-4333-adde-31efeaa36e77)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"56c6b6ac2ca7174695af317f44ecdf75f71cc877eafa526461a583b8c56a0587\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-9f7667bb8-bz8j9" podUID="64f79500-120a-4333-adde-31efeaa36e77" Mar 6 02:39:42.631744 containerd[1569]: time="2026-03-06T02:39:42.628549552Z" level=info msg="StartContainer for \"b3a703a6ccaa7a44ddca005f1c20cecd9636e807b4d8a348944b76b31df5487b\" returns successfully" Mar 6 02:39:42.642678 containerd[1569]: time="2026-03-06T02:39:42.638466870Z" level=error msg="Failed to destroy network for sandbox \"e98ff7ce1adf25ffb57bef779c189c4a7a294c1798bb85fda90305558e6acb7b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.651743 containerd[1569]: time="2026-03-06T02:39:42.651689374Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-gqfqd,Uid:74be49b3-6c58-4ae9-86a2-c8139f48d9c9,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e98ff7ce1adf25ffb57bef779c189c4a7a294c1798bb85fda90305558e6acb7b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.653780 kubelet[2879]: E0306 02:39:42.653737 2879 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e98ff7ce1adf25ffb57bef779c189c4a7a294c1798bb85fda90305558e6acb7b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 6 02:39:42.655670 kubelet[2879]: E0306 02:39:42.655638 2879 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e98ff7ce1adf25ffb57bef779c189c4a7a294c1798bb85fda90305558e6acb7b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-gqfqd" Mar 6 02:39:42.655815 kubelet[2879]: E0306 02:39:42.655785 2879 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e98ff7ce1adf25ffb57bef779c189c4a7a294c1798bb85fda90305558e6acb7b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-gqfqd" Mar 6 02:39:42.655988 kubelet[2879]: E0306 02:39:42.655953 2879 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-gqfqd_calico-system(74be49b3-6c58-4ae9-86a2-c8139f48d9c9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-gqfqd_calico-system(74be49b3-6c58-4ae9-86a2-c8139f48d9c9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e98ff7ce1adf25ffb57bef779c189c4a7a294c1798bb85fda90305558e6acb7b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-gqfqd" podUID="74be49b3-6c58-4ae9-86a2-c8139f48d9c9" Mar 6 02:39:43.193565 systemd[1]: run-netns-cni\x2d081ec321\x2ddf32\x2d3048\x2d63d1\x2d76c16c6db0e3.mount: Deactivated successfully. Mar 6 02:39:43.193730 systemd[1]: run-netns-cni\x2da4e446fd\x2d81eb\x2d3586\x2db462\x2d819a78f4ffa3.mount: Deactivated successfully. Mar 6 02:39:43.193808 systemd[1]: run-netns-cni\x2d2cbcb818\x2d2c88\x2dc8e0\x2dd303\x2dd7a7667cd59e.mount: Deactivated successfully. Mar 6 02:39:43.193872 systemd[1]: run-netns-cni\x2db0cf0704\x2d0778\x2d2acc\x2d1909\x2d8df67a4edec2.mount: Deactivated successfully. Mar 6 02:39:43.645860 kubelet[2879]: I0306 02:39:43.643916 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-node-cvqkx" podStartSLOduration=4.411062703 podStartE2EDuration="1m9.643897782s" podCreationTimestamp="2026-03-06 02:38:34 +0000 UTC" firstStartedPulling="2026-03-06 02:38:36.149790016 +0000 UTC m=+52.156988509" lastFinishedPulling="2026-03-06 02:39:41.382625096 +0000 UTC m=+117.389823588" observedRunningTime="2026-03-06 02:39:43.640047003 +0000 UTC m=+119.647245495" watchObservedRunningTime="2026-03-06 02:39:43.643897782 +0000 UTC m=+119.651096273" Mar 6 02:39:44.744686 kubelet[2879]: I0306 02:39:44.742717 2879 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-ca-bundle\") pod \"ee2caea9-f1f8-4884-8243-566cb757b833\" (UID: \"ee2caea9-f1f8-4884-8243-566cb757b833\") " Mar 6 02:39:44.748730 kubelet[2879]: I0306 02:39:44.748085 2879 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-nginx-config" pod "ee2caea9-f1f8-4884-8243-566cb757b833" (UID: "ee2caea9-f1f8-4884-8243-566cb757b833"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 6 02:39:44.756624 kubelet[2879]: I0306 02:39:44.752018 2879 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-nginx-config\" (UniqueName: \"kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-nginx-config\") pod \"ee2caea9-f1f8-4884-8243-566cb757b833\" (UID: \"ee2caea9-f1f8-4884-8243-566cb757b833\") " Mar 6 02:39:44.762559 kubelet[2879]: I0306 02:39:44.761607 2879 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/projected/ee2caea9-f1f8-4884-8243-566cb757b833-kube-api-access-gvhjv\" (UniqueName: \"kubernetes.io/projected/ee2caea9-f1f8-4884-8243-566cb757b833-kube-api-access-gvhjv\") pod \"ee2caea9-f1f8-4884-8243-566cb757b833\" (UID: \"ee2caea9-f1f8-4884-8243-566cb757b833\") " Mar 6 02:39:44.769508 kubelet[2879]: I0306 02:39:44.752826 2879 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-ca-bundle" pod "ee2caea9-f1f8-4884-8243-566cb757b833" (UID: "ee2caea9-f1f8-4884-8243-566cb757b833"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 6 02:39:44.776872 kubelet[2879]: I0306 02:39:44.776644 2879 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/secret/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-backend-key-pair\") pod \"ee2caea9-f1f8-4884-8243-566cb757b833\" (UID: \"ee2caea9-f1f8-4884-8243-566cb757b833\") " Mar 6 02:39:44.777698 kubelet[2879]: I0306 02:39:44.777046 2879 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Mar 6 02:39:44.778924 kubelet[2879]: I0306 02:39:44.778742 2879 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/ee2caea9-f1f8-4884-8243-566cb757b833-nginx-config\") on node \"localhost\" DevicePath \"\"" Mar 6 02:39:44.786629 systemd[1]: var-lib-kubelet-pods-ee2caea9\x2df1f8\x2d4884\x2d8243\x2d566cb757b833-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dgvhjv.mount: Deactivated successfully. Mar 6 02:39:44.796589 kubelet[2879]: I0306 02:39:44.794627 2879 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2caea9-f1f8-4884-8243-566cb757b833-kube-api-access-gvhjv" pod "ee2caea9-f1f8-4884-8243-566cb757b833" (UID: "ee2caea9-f1f8-4884-8243-566cb757b833"). InnerVolumeSpecName "kube-api-access-gvhjv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 6 02:39:44.801821 systemd[1]: var-lib-kubelet-pods-ee2caea9\x2df1f8\x2d4884\x2d8243\x2d566cb757b833-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Mar 6 02:39:44.808092 kubelet[2879]: I0306 02:39:44.805997 2879 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-backend-key-pair" pod "ee2caea9-f1f8-4884-8243-566cb757b833" (UID: "ee2caea9-f1f8-4884-8243-566cb757b833"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 6 02:39:44.880667 kubelet[2879]: I0306 02:39:44.880129 2879 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ee2caea9-f1f8-4884-8243-566cb757b833-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Mar 6 02:39:44.880667 kubelet[2879]: I0306 02:39:44.880528 2879 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-gvhjv\" (UniqueName: \"kubernetes.io/projected/ee2caea9-f1f8-4884-8243-566cb757b833-kube-api-access-gvhjv\") on node \"localhost\" DevicePath \"\"" Mar 6 02:39:45.595808 systemd[1]: Removed slice kubepods-besteffort-podee2caea9_f1f8_4884_8243_566cb757b833.slice - libcontainer container kubepods-besteffort-podee2caea9_f1f8_4884_8243_566cb757b833.slice. Mar 6 02:39:46.287019 kubelet[2879]: I0306 02:39:46.286925 2879 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="ee2caea9-f1f8-4884-8243-566cb757b833" path="/var/lib/kubelet/pods/ee2caea9-f1f8-4884-8243-566cb757b833/volumes" Mar 6 02:39:48.756148 systemd[1]: Created slice kubepods-besteffort-pod66dfdc98_74f3_4a6b_8c9f_2ef6621fd9fa.slice - libcontainer container kubepods-besteffort-pod66dfdc98_74f3_4a6b_8c9f_2ef6621fd9fa.slice. Mar 6 02:39:48.779936 kubelet[2879]: I0306 02:39:48.779767 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa-whisker-ca-bundle\") pod \"whisker-57c85f46f-59gzc\" (UID: \"66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa\") " pod="calico-system/whisker-57c85f46f-59gzc" Mar 6 02:39:48.779936 kubelet[2879]: I0306 02:39:48.779886 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa-whisker-backend-key-pair\") pod \"whisker-57c85f46f-59gzc\" (UID: \"66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa\") " pod="calico-system/whisker-57c85f46f-59gzc" Mar 6 02:39:48.780566 kubelet[2879]: I0306 02:39:48.779938 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfc6g\" (UniqueName: \"kubernetes.io/projected/66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa-kube-api-access-mfc6g\") pod \"whisker-57c85f46f-59gzc\" (UID: \"66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa\") " pod="calico-system/whisker-57c85f46f-59gzc" Mar 6 02:39:48.780566 kubelet[2879]: I0306 02:39:48.779976 2879 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa-nginx-config\") pod \"whisker-57c85f46f-59gzc\" (UID: \"66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa\") " pod="calico-system/whisker-57c85f46f-59gzc" Mar 6 02:39:49.085525 containerd[1569]: time="2026-03-06T02:39:49.084725038Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-57c85f46f-59gzc,Uid:66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:49.888887 systemd-networkd[1497]: cali4e266a44a6e: Link UP Mar 6 02:39:49.892532 systemd-networkd[1497]: cali4e266a44a6e: Gained carrier Mar 6 02:39:49.962779 containerd[1569]: 2026-03-06 02:39:49.232 [ERROR][4162] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Mar 6 02:39:49.962779 containerd[1569]: 2026-03-06 02:39:49.372 [INFO][4162] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--57c85f46f--59gzc-eth0 whisker-57c85f46f- calico-system 66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa 1110 0 2026-03-06 02:39:46 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:57c85f46f projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-57c85f46f-59gzc eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali4e266a44a6e [] [] }} ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Namespace="calico-system" Pod="whisker-57c85f46f-59gzc" WorkloadEndpoint="localhost-k8s-whisker--57c85f46f--59gzc-" Mar 6 02:39:49.962779 containerd[1569]: 2026-03-06 02:39:49.372 [INFO][4162] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Namespace="calico-system" Pod="whisker-57c85f46f-59gzc" WorkloadEndpoint="localhost-k8s-whisker--57c85f46f--59gzc-eth0" Mar 6 02:39:49.962779 containerd[1569]: 2026-03-06 02:39:49.480 [INFO][4175] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" HandleID="k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Workload="localhost-k8s-whisker--57c85f46f--59gzc-eth0" Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.535 [INFO][4175] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" HandleID="k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Workload="localhost-k8s-whisker--57c85f46f--59gzc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001b21f0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-57c85f46f-59gzc", "timestamp":"2026-03-06 02:39:49.480007525 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000156dc0)} Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.535 [INFO][4175] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.536 [INFO][4175] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.536 [INFO][4175] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.556 [INFO][4175] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" host="localhost" Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.679 [INFO][4175] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.735 [INFO][4175] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.747 [INFO][4175] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.757 [INFO][4175] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:49.964540 containerd[1569]: 2026-03-06 02:39:49.757 [INFO][4175] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" host="localhost" Mar 6 02:39:49.965051 containerd[1569]: 2026-03-06 02:39:49.763 [INFO][4175] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec Mar 6 02:39:49.965051 containerd[1569]: 2026-03-06 02:39:49.779 [INFO][4175] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" host="localhost" Mar 6 02:39:49.965051 containerd[1569]: 2026-03-06 02:39:49.837 [INFO][4175] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" host="localhost" Mar 6 02:39:49.965051 containerd[1569]: 2026-03-06 02:39:49.837 [INFO][4175] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" host="localhost" Mar 6 02:39:49.965051 containerd[1569]: 2026-03-06 02:39:49.838 [INFO][4175] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 6 02:39:49.965051 containerd[1569]: 2026-03-06 02:39:49.838 [INFO][4175] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" HandleID="k8s-pod-network.33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Workload="localhost-k8s-whisker--57c85f46f--59gzc-eth0" Mar 6 02:39:49.965467 containerd[1569]: 2026-03-06 02:39:49.849 [INFO][4162] cni-plugin/k8s.go 418: Populated endpoint ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Namespace="calico-system" Pod="whisker-57c85f46f-59gzc" WorkloadEndpoint="localhost-k8s-whisker--57c85f46f--59gzc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--57c85f46f--59gzc-eth0", GenerateName:"whisker-57c85f46f-", Namespace:"calico-system", SelfLink:"", UID:"66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa", ResourceVersion:"1110", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 39, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"57c85f46f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-57c85f46f-59gzc", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali4e266a44a6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:49.965467 containerd[1569]: 2026-03-06 02:39:49.849 [INFO][4162] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Namespace="calico-system" Pod="whisker-57c85f46f-59gzc" WorkloadEndpoint="localhost-k8s-whisker--57c85f46f--59gzc-eth0" Mar 6 02:39:49.965678 containerd[1569]: 2026-03-06 02:39:49.849 [INFO][4162] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4e266a44a6e ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Namespace="calico-system" Pod="whisker-57c85f46f-59gzc" WorkloadEndpoint="localhost-k8s-whisker--57c85f46f--59gzc-eth0" Mar 6 02:39:49.965678 containerd[1569]: 2026-03-06 02:39:49.895 [INFO][4162] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Namespace="calico-system" Pod="whisker-57c85f46f-59gzc" WorkloadEndpoint="localhost-k8s-whisker--57c85f46f--59gzc-eth0" Mar 6 02:39:49.965741 containerd[1569]: 2026-03-06 02:39:49.898 [INFO][4162] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Namespace="calico-system" Pod="whisker-57c85f46f-59gzc" WorkloadEndpoint="localhost-k8s-whisker--57c85f46f--59gzc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--57c85f46f--59gzc-eth0", GenerateName:"whisker-57c85f46f-", Namespace:"calico-system", SelfLink:"", UID:"66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa", ResourceVersion:"1110", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 39, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"57c85f46f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec", Pod:"whisker-57c85f46f-59gzc", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali4e266a44a6e", MAC:"da:58:65:0c:12:a7", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:49.965904 containerd[1569]: 2026-03-06 02:39:49.952 [INFO][4162] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" Namespace="calico-system" Pod="whisker-57c85f46f-59gzc" WorkloadEndpoint="localhost-k8s-whisker--57c85f46f--59gzc-eth0" Mar 6 02:39:50.221463 containerd[1569]: time="2026-03-06T02:39:50.219871465Z" level=info msg="connecting to shim 33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec" address="unix:///run/containerd/s/89ca6d057265e175a1c6ca02d16930f20d9e13cd06060d9647f0078f2afc4bf8" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:39:50.365034 systemd[1]: Started cri-containerd-33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec.scope - libcontainer container 33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec. Mar 6 02:39:50.460149 systemd-resolved[1399]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 02:39:50.643001 containerd[1569]: time="2026-03-06T02:39:50.642910156Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-57c85f46f-59gzc,Uid:66dfdc98-74f3-4a6b-8c9f-2ef6621fd9fa,Namespace:calico-system,Attempt:0,} returns sandbox id \"33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec\"" Mar 6 02:39:50.666594 containerd[1569]: time="2026-03-06T02:39:50.666533897Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\"" Mar 6 02:39:50.993800 systemd-networkd[1497]: cali4e266a44a6e: Gained IPv6LL Mar 6 02:39:52.227115 containerd[1569]: time="2026-03-06T02:39:52.226204139Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:52.239552 containerd[1569]: time="2026-03-06T02:39:52.238985641Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.4: active requests=0, bytes read=6039889" Mar 6 02:39:52.249390 containerd[1569]: time="2026-03-06T02:39:52.248644277Z" level=info msg="ImageCreate event name:\"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:52.267805 containerd[1569]: time="2026-03-06T02:39:52.267547885Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:52.274207 containerd[1569]: time="2026-03-06T02:39:52.273933653Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.4\" with image id \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\", size \"7595926\" in 1.607231502s" Mar 6 02:39:52.274207 containerd[1569]: time="2026-03-06T02:39:52.274115063Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\" returns image reference \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\"" Mar 6 02:39:52.290794 containerd[1569]: time="2026-03-06T02:39:52.290745231Z" level=info msg="CreateContainer within sandbox \"33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Mar 6 02:39:52.357738 containerd[1569]: time="2026-03-06T02:39:52.357589539Z" level=info msg="Container 653b0693ee559715d47bdeac5b99f805b4702105d04d2c877bdf7227e6a8954f: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:39:52.364781 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1800303828.mount: Deactivated successfully. Mar 6 02:39:52.406760 containerd[1569]: time="2026-03-06T02:39:52.404919118Z" level=info msg="CreateContainer within sandbox \"33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"653b0693ee559715d47bdeac5b99f805b4702105d04d2c877bdf7227e6a8954f\"" Mar 6 02:39:52.415102 containerd[1569]: time="2026-03-06T02:39:52.415060899Z" level=info msg="StartContainer for \"653b0693ee559715d47bdeac5b99f805b4702105d04d2c877bdf7227e6a8954f\"" Mar 6 02:39:52.419744 containerd[1569]: time="2026-03-06T02:39:52.419568285Z" level=info msg="connecting to shim 653b0693ee559715d47bdeac5b99f805b4702105d04d2c877bdf7227e6a8954f" address="unix:///run/containerd/s/89ca6d057265e175a1c6ca02d16930f20d9e13cd06060d9647f0078f2afc4bf8" protocol=ttrpc version=3 Mar 6 02:39:52.526904 systemd[1]: Started cri-containerd-653b0693ee559715d47bdeac5b99f805b4702105d04d2c877bdf7227e6a8954f.scope - libcontainer container 653b0693ee559715d47bdeac5b99f805b4702105d04d2c877bdf7227e6a8954f. Mar 6 02:39:52.669403 containerd[1569]: time="2026-03-06T02:39:52.669049923Z" level=info msg="StartContainer for \"653b0693ee559715d47bdeac5b99f805b4702105d04d2c877bdf7227e6a8954f\" returns successfully" Mar 6 02:39:52.677818 containerd[1569]: time="2026-03-06T02:39:52.677729917Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\"" Mar 6 02:39:53.078221 systemd-networkd[1497]: vxlan.calico: Link UP Mar 6 02:39:53.079176 systemd-networkd[1497]: vxlan.calico: Gained carrier Mar 6 02:39:53.287625 containerd[1569]: time="2026-03-06T02:39:53.287503915Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-d54f5f4cf-vrw2m,Uid:be2e66c8-0f54-46c3-af38-09952dbbac49,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:53.985414 systemd-networkd[1497]: calieb70baec39d: Link UP Mar 6 02:39:54.018404 systemd-networkd[1497]: calieb70baec39d: Gained carrier Mar 6 02:39:54.146019 containerd[1569]: 2026-03-06 02:39:53.583 [INFO][4462] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0 calico-apiserver-d54f5f4cf- calico-system be2e66c8-0f54-46c3-af38-09952dbbac49 1045 0 2026-03-06 02:38:32 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:d54f5f4cf projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-d54f5f4cf-vrw2m eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] calieb70baec39d [] [] }} ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-vrw2m" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-" Mar 6 02:39:54.146019 containerd[1569]: 2026-03-06 02:39:53.587 [INFO][4462] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-vrw2m" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" Mar 6 02:39:54.146019 containerd[1569]: 2026-03-06 02:39:53.706 [INFO][4476] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" HandleID="k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Workload="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.751 [INFO][4476] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" HandleID="k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Workload="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0004fba80), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-apiserver-d54f5f4cf-vrw2m", "timestamp":"2026-03-06 02:39:53.706608853 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0000d8580)} Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.751 [INFO][4476] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.751 [INFO][4476] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.751 [INFO][4476] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.762 [INFO][4476] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" host="localhost" Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.795 [INFO][4476] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.858 [INFO][4476] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.870 [INFO][4476] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.885 [INFO][4476] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:54.146632 containerd[1569]: 2026-03-06 02:39:53.885 [INFO][4476] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" host="localhost" Mar 6 02:39:54.147749 containerd[1569]: 2026-03-06 02:39:53.893 [INFO][4476] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d Mar 6 02:39:54.147749 containerd[1569]: 2026-03-06 02:39:53.935 [INFO][4476] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" host="localhost" Mar 6 02:39:54.147749 containerd[1569]: 2026-03-06 02:39:53.962 [INFO][4476] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" host="localhost" Mar 6 02:39:54.147749 containerd[1569]: 2026-03-06 02:39:53.964 [INFO][4476] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" host="localhost" Mar 6 02:39:54.147749 containerd[1569]: 2026-03-06 02:39:53.964 [INFO][4476] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 6 02:39:54.147749 containerd[1569]: 2026-03-06 02:39:53.964 [INFO][4476] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" HandleID="k8s-pod-network.4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Workload="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" Mar 6 02:39:54.148519 containerd[1569]: 2026-03-06 02:39:53.971 [INFO][4462] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-vrw2m" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0", GenerateName:"calico-apiserver-d54f5f4cf-", Namespace:"calico-system", SelfLink:"", UID:"be2e66c8-0f54-46c3-af38-09952dbbac49", ResourceVersion:"1045", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"d54f5f4cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-d54f5f4cf-vrw2m", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calieb70baec39d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:54.148725 containerd[1569]: 2026-03-06 02:39:53.971 [INFO][4462] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-vrw2m" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" Mar 6 02:39:54.148725 containerd[1569]: 2026-03-06 02:39:53.971 [INFO][4462] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieb70baec39d ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-vrw2m" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" Mar 6 02:39:54.148725 containerd[1569]: 2026-03-06 02:39:54.032 [INFO][4462] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-vrw2m" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" Mar 6 02:39:54.148790 containerd[1569]: 2026-03-06 02:39:54.038 [INFO][4462] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-vrw2m" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0", GenerateName:"calico-apiserver-d54f5f4cf-", Namespace:"calico-system", SelfLink:"", UID:"be2e66c8-0f54-46c3-af38-09952dbbac49", ResourceVersion:"1045", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"d54f5f4cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d", Pod:"calico-apiserver-d54f5f4cf-vrw2m", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calieb70baec39d", MAC:"1a:dc:f2:b0:e1:18", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:54.148961 containerd[1569]: 2026-03-06 02:39:54.072 [INFO][4462] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-vrw2m" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--vrw2m-eth0" Mar 6 02:39:54.282627 containerd[1569]: time="2026-03-06T02:39:54.280593844Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5cf8bb7c94-m8khs,Uid:2776157f-9f5c-4064-8095-991904ea8002,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:54.284997 containerd[1569]: time="2026-03-06T02:39:54.284925828Z" level=info msg="connecting to shim 4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d" address="unix:///run/containerd/s/808e894657a328f829b0e8819ca1522ddcf88fe34e7a257dbb4241b470dae5b1" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:39:54.290144 containerd[1569]: time="2026-03-06T02:39:54.289723766Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-gqfqd,Uid:74be49b3-6c58-4ae9-86a2-c8139f48d9c9,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:54.394006 systemd[1]: Started cri-containerd-4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d.scope - libcontainer container 4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d. Mar 6 02:39:54.441766 systemd-networkd[1497]: vxlan.calico: Gained IPv6LL Mar 6 02:39:54.501749 systemd-resolved[1399]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 02:39:54.731159 containerd[1569]: time="2026-03-06T02:39:54.730933460Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-d54f5f4cf-vrw2m,Uid:be2e66c8-0f54-46c3-af38-09952dbbac49,Namespace:calico-system,Attempt:0,} returns sandbox id \"4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d\"" Mar 6 02:39:54.831563 systemd-networkd[1497]: cali86802aad6d5: Link UP Mar 6 02:39:54.833648 systemd-networkd[1497]: cali86802aad6d5: Gained carrier Mar 6 02:39:54.916205 containerd[1569]: 2026-03-06 02:39:54.463 [INFO][4575] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--gqfqd-eth0 csi-node-driver- calico-system 74be49b3-6c58-4ae9-86a2-c8139f48d9c9 814 0 2026-03-06 02:38:35 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:589b8b8d94 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-gqfqd eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali86802aad6d5 [] [] }} ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Namespace="calico-system" Pod="csi-node-driver-gqfqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--gqfqd-" Mar 6 02:39:54.916205 containerd[1569]: 2026-03-06 02:39:54.464 [INFO][4575] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Namespace="calico-system" Pod="csi-node-driver-gqfqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--gqfqd-eth0" Mar 6 02:39:54.916205 containerd[1569]: 2026-03-06 02:39:54.586 [INFO][4616] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" HandleID="k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Workload="localhost-k8s-csi--node--driver--gqfqd-eth0" Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.627 [INFO][4616] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" HandleID="k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Workload="localhost-k8s-csi--node--driver--gqfqd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ad510), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-gqfqd", "timestamp":"2026-03-06 02:39:54.586828409 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0004682c0)} Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.628 [INFO][4616] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.628 [INFO][4616] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.628 [INFO][4616] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.635 [INFO][4616] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" host="localhost" Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.659 [INFO][4616] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.698 [INFO][4616] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.731 [INFO][4616] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.741 [INFO][4616] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:54.919785 containerd[1569]: 2026-03-06 02:39:54.742 [INFO][4616] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" host="localhost" Mar 6 02:39:54.922077 containerd[1569]: 2026-03-06 02:39:54.757 [INFO][4616] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889 Mar 6 02:39:54.922077 containerd[1569]: 2026-03-06 02:39:54.770 [INFO][4616] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" host="localhost" Mar 6 02:39:54.922077 containerd[1569]: 2026-03-06 02:39:54.810 [INFO][4616] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" host="localhost" Mar 6 02:39:54.922077 containerd[1569]: 2026-03-06 02:39:54.810 [INFO][4616] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" host="localhost" Mar 6 02:39:54.922077 containerd[1569]: 2026-03-06 02:39:54.810 [INFO][4616] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 6 02:39:54.922077 containerd[1569]: 2026-03-06 02:39:54.811 [INFO][4616] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" HandleID="k8s-pod-network.ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Workload="localhost-k8s-csi--node--driver--gqfqd-eth0" Mar 6 02:39:54.923522 containerd[1569]: 2026-03-06 02:39:54.820 [INFO][4575] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Namespace="calico-system" Pod="csi-node-driver-gqfqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--gqfqd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--gqfqd-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"74be49b3-6c58-4ae9-86a2-c8139f48d9c9", ResourceVersion:"814", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"589b8b8d94", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-gqfqd", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali86802aad6d5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:54.923713 containerd[1569]: 2026-03-06 02:39:54.820 [INFO][4575] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Namespace="calico-system" Pod="csi-node-driver-gqfqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--gqfqd-eth0" Mar 6 02:39:54.923713 containerd[1569]: 2026-03-06 02:39:54.820 [INFO][4575] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali86802aad6d5 ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Namespace="calico-system" Pod="csi-node-driver-gqfqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--gqfqd-eth0" Mar 6 02:39:54.923713 containerd[1569]: 2026-03-06 02:39:54.834 [INFO][4575] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Namespace="calico-system" Pod="csi-node-driver-gqfqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--gqfqd-eth0" Mar 6 02:39:54.923834 containerd[1569]: 2026-03-06 02:39:54.845 [INFO][4575] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Namespace="calico-system" Pod="csi-node-driver-gqfqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--gqfqd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--gqfqd-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"74be49b3-6c58-4ae9-86a2-c8139f48d9c9", ResourceVersion:"814", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"589b8b8d94", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889", Pod:"csi-node-driver-gqfqd", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali86802aad6d5", MAC:"92:ff:ef:33:63:2d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:54.924007 containerd[1569]: 2026-03-06 02:39:54.889 [INFO][4575] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" Namespace="calico-system" Pod="csi-node-driver-gqfqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--gqfqd-eth0" Mar 6 02:39:55.068212 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2099966065.mount: Deactivated successfully. Mar 6 02:39:55.083391 containerd[1569]: time="2026-03-06T02:39:55.083201829Z" level=info msg="connecting to shim ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889" address="unix:///run/containerd/s/704dd4bf980acf83e2286c33e0dfa7a716b21035f8f6a97e47896491667cd04d" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:39:55.105566 systemd-networkd[1497]: calid5572191cf2: Link UP Mar 6 02:39:55.106834 systemd-networkd[1497]: calid5572191cf2: Gained carrier Mar 6 02:39:55.196222 containerd[1569]: time="2026-03-06T02:39:55.195750901Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:55.200202 containerd[1569]: time="2026-03-06T02:39:55.200135782Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.4: active requests=0, bytes read=17609475" Mar 6 02:39:55.211899 containerd[1569]: time="2026-03-06T02:39:55.211805981Z" level=info msg="ImageCreate event name:\"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:55.227207 containerd[1569]: time="2026-03-06T02:39:55.225741723Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" with image id \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\", size \"17609305\" in 2.547927649s" Mar 6 02:39:55.228191 containerd[1569]: time="2026-03-06T02:39:55.227800246Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" returns image reference \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\"" Mar 6 02:39:55.228191 containerd[1569]: time="2026-03-06T02:39:55.227889446Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:39:55.260456 containerd[1569]: time="2026-03-06T02:39:55.260163124Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Mar 6 02:39:55.277883 containerd[1569]: 2026-03-06 02:39:54.540 [INFO][4567] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0 calico-kube-controllers-5cf8bb7c94- calico-system 2776157f-9f5c-4064-8095-991904ea8002 1047 0 2026-03-06 02:38:35 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5cf8bb7c94 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-5cf8bb7c94-m8khs eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calid5572191cf2 [] [] }} ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Namespace="calico-system" Pod="calico-kube-controllers-5cf8bb7c94-m8khs" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-" Mar 6 02:39:55.277883 containerd[1569]: 2026-03-06 02:39:54.544 [INFO][4567] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Namespace="calico-system" Pod="calico-kube-controllers-5cf8bb7c94-m8khs" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" Mar 6 02:39:55.277883 containerd[1569]: 2026-03-06 02:39:54.684 [INFO][4626] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" HandleID="k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Workload="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.729 [INFO][4626] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" HandleID="k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Workload="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000276b20), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-5cf8bb7c94-m8khs", "timestamp":"2026-03-06 02:39:54.684564956 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0000fe840)} Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.729 [INFO][4626] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.812 [INFO][4626] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.814 [INFO][4626] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.827 [INFO][4626] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" host="localhost" Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.855 [INFO][4626] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.935 [INFO][4626] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.962 [INFO][4626] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:55.278409 containerd[1569]: 2026-03-06 02:39:54.972 [INFO][4626] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:55.278883 containerd[1569]: 2026-03-06 02:39:54.972 [INFO][4626] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" host="localhost" Mar 6 02:39:55.278883 containerd[1569]: 2026-03-06 02:39:54.986 [INFO][4626] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875 Mar 6 02:39:55.278883 containerd[1569]: 2026-03-06 02:39:55.036 [INFO][4626] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" host="localhost" Mar 6 02:39:55.278883 containerd[1569]: 2026-03-06 02:39:55.072 [INFO][4626] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" host="localhost" Mar 6 02:39:55.278883 containerd[1569]: 2026-03-06 02:39:55.076 [INFO][4626] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" host="localhost" Mar 6 02:39:55.278883 containerd[1569]: 2026-03-06 02:39:55.076 [INFO][4626] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 6 02:39:55.278883 containerd[1569]: 2026-03-06 02:39:55.076 [INFO][4626] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" HandleID="k8s-pod-network.db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Workload="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" Mar 6 02:39:55.279113 containerd[1569]: 2026-03-06 02:39:55.095 [INFO][4567] cni-plugin/k8s.go 418: Populated endpoint ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Namespace="calico-system" Pod="calico-kube-controllers-5cf8bb7c94-m8khs" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0", GenerateName:"calico-kube-controllers-5cf8bb7c94-", Namespace:"calico-system", SelfLink:"", UID:"2776157f-9f5c-4064-8095-991904ea8002", ResourceVersion:"1047", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5cf8bb7c94", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-5cf8bb7c94-m8khs", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calid5572191cf2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:55.279571 containerd[1569]: 2026-03-06 02:39:55.096 [INFO][4567] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Namespace="calico-system" Pod="calico-kube-controllers-5cf8bb7c94-m8khs" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" Mar 6 02:39:55.279571 containerd[1569]: 2026-03-06 02:39:55.096 [INFO][4567] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid5572191cf2 ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Namespace="calico-system" Pod="calico-kube-controllers-5cf8bb7c94-m8khs" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" Mar 6 02:39:55.279571 containerd[1569]: 2026-03-06 02:39:55.113 [INFO][4567] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Namespace="calico-system" Pod="calico-kube-controllers-5cf8bb7c94-m8khs" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" Mar 6 02:39:55.279695 containerd[1569]: 2026-03-06 02:39:55.124 [INFO][4567] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Namespace="calico-system" Pod="calico-kube-controllers-5cf8bb7c94-m8khs" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0", GenerateName:"calico-kube-controllers-5cf8bb7c94-", Namespace:"calico-system", SelfLink:"", UID:"2776157f-9f5c-4064-8095-991904ea8002", ResourceVersion:"1047", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5cf8bb7c94", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875", Pod:"calico-kube-controllers-5cf8bb7c94-m8khs", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calid5572191cf2", MAC:"ae:34:42:76:9d:71", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:55.280110 containerd[1569]: 2026-03-06 02:39:55.232 [INFO][4567] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" Namespace="calico-system" Pod="calico-kube-controllers-5cf8bb7c94-m8khs" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5cf8bb7c94--m8khs-eth0" Mar 6 02:39:55.282775 containerd[1569]: time="2026-03-06T02:39:55.282733988Z" level=info msg="CreateContainer within sandbox \"33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Mar 6 02:39:55.284466 containerd[1569]: time="2026-03-06T02:39:55.283492767Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-rghxd,Uid:48337397-07a0-44d5-b732-338087eee7ba,Namespace:kube-system,Attempt:0,}" Mar 6 02:39:55.351071 systemd[1]: Started cri-containerd-ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889.scope - libcontainer container ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889. Mar 6 02:39:55.368988 containerd[1569]: time="2026-03-06T02:39:55.368884949Z" level=info msg="Container 71ae7293124bc6503b694838d260c04f99b86193d6217db40b3e38ddb33c3a10: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:39:55.448711 containerd[1569]: time="2026-03-06T02:39:55.448620575Z" level=info msg="CreateContainer within sandbox \"33e49197b8c08e0574d5928a406512b1d23d19daee0c4ee63810f2344c38b7ec\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"71ae7293124bc6503b694838d260c04f99b86193d6217db40b3e38ddb33c3a10\"" Mar 6 02:39:55.454129 containerd[1569]: time="2026-03-06T02:39:55.454082077Z" level=info msg="StartContainer for \"71ae7293124bc6503b694838d260c04f99b86193d6217db40b3e38ddb33c3a10\"" Mar 6 02:39:55.462113 containerd[1569]: time="2026-03-06T02:39:55.462071428Z" level=info msg="connecting to shim 71ae7293124bc6503b694838d260c04f99b86193d6217db40b3e38ddb33c3a10" address="unix:///run/containerd/s/89ca6d057265e175a1c6ca02d16930f20d9e13cd06060d9647f0078f2afc4bf8" protocol=ttrpc version=3 Mar 6 02:39:55.502069 systemd-resolved[1399]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 02:39:55.554706 systemd[1]: Started cri-containerd-71ae7293124bc6503b694838d260c04f99b86193d6217db40b3e38ddb33c3a10.scope - libcontainer container 71ae7293124bc6503b694838d260c04f99b86193d6217db40b3e38ddb33c3a10. Mar 6 02:39:55.576639 containerd[1569]: time="2026-03-06T02:39:55.576470975Z" level=info msg="connecting to shim db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875" address="unix:///run/containerd/s/df6d070569a4d3df533c26aa7e25933ab104d88917e7e55ee0eee91410ea798a" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:39:55.802087 systemd[1]: Started cri-containerd-db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875.scope - libcontainer container db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875. Mar 6 02:39:55.804902 containerd[1569]: time="2026-03-06T02:39:55.804850457Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-gqfqd,Uid:74be49b3-6c58-4ae9-86a2-c8139f48d9c9,Namespace:calico-system,Attempt:0,} returns sandbox id \"ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889\"" Mar 6 02:39:55.916759 systemd-resolved[1399]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 02:39:55.981661 systemd-networkd[1497]: calieb70baec39d: Gained IPv6LL Mar 6 02:39:56.048100 containerd[1569]: time="2026-03-06T02:39:56.048052905Z" level=info msg="StartContainer for \"71ae7293124bc6503b694838d260c04f99b86193d6217db40b3e38ddb33c3a10\" returns successfully" Mar 6 02:39:56.143551 systemd-networkd[1497]: calie188408e6a6: Link UP Mar 6 02:39:56.143906 systemd-networkd[1497]: calie188408e6a6: Gained carrier Mar 6 02:39:56.170088 systemd-networkd[1497]: cali86802aad6d5: Gained IPv6LL Mar 6 02:39:56.244214 containerd[1569]: 2026-03-06 02:39:55.567 [INFO][4705] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7d764666f9--rghxd-eth0 coredns-7d764666f9- kube-system 48337397-07a0-44d5-b732-338087eee7ba 1036 0 2026-03-06 02:37:45 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7d764666f9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7d764666f9-rghxd eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calie188408e6a6 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Namespace="kube-system" Pod="coredns-7d764666f9-rghxd" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--rghxd-" Mar 6 02:39:56.244214 containerd[1569]: 2026-03-06 02:39:55.567 [INFO][4705] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Namespace="kube-system" Pod="coredns-7d764666f9-rghxd" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--rghxd-eth0" Mar 6 02:39:56.244214 containerd[1569]: 2026-03-06 02:39:55.821 [INFO][4767] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" HandleID="k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Workload="localhost-k8s-coredns--7d764666f9--rghxd-eth0" Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:55.842 [INFO][4767] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" HandleID="k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Workload="localhost-k8s-coredns--7d764666f9--rghxd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0004d18c0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7d764666f9-rghxd", "timestamp":"2026-03-06 02:39:55.821590997 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000512420)} Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:55.846 [INFO][4767] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:55.846 [INFO][4767] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:55.846 [INFO][4767] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:55.869 [INFO][4767] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" host="localhost" Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:55.942 [INFO][4767] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:55.987 [INFO][4767] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:55.998 [INFO][4767] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:56.007 [INFO][4767] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:56.245008 containerd[1569]: 2026-03-06 02:39:56.007 [INFO][4767] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" host="localhost" Mar 6 02:39:56.245955 containerd[1569]: 2026-03-06 02:39:56.016 [INFO][4767] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f Mar 6 02:39:56.245955 containerd[1569]: 2026-03-06 02:39:56.043 [INFO][4767] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" host="localhost" Mar 6 02:39:56.245955 containerd[1569]: 2026-03-06 02:39:56.093 [INFO][4767] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" host="localhost" Mar 6 02:39:56.245955 containerd[1569]: 2026-03-06 02:39:56.093 [INFO][4767] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" host="localhost" Mar 6 02:39:56.245955 containerd[1569]: 2026-03-06 02:39:56.093 [INFO][4767] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 6 02:39:56.245955 containerd[1569]: 2026-03-06 02:39:56.093 [INFO][4767] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" HandleID="k8s-pod-network.2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Workload="localhost-k8s-coredns--7d764666f9--rghxd-eth0" Mar 6 02:39:56.246146 containerd[1569]: 2026-03-06 02:39:56.121 [INFO][4705] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Namespace="kube-system" Pod="coredns-7d764666f9-rghxd" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--rghxd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--rghxd-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"48337397-07a0-44d5-b732-338087eee7ba", ResourceVersion:"1036", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 37, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7d764666f9-rghxd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calie188408e6a6", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:56.246146 containerd[1569]: 2026-03-06 02:39:56.122 [INFO][4705] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Namespace="kube-system" Pod="coredns-7d764666f9-rghxd" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--rghxd-eth0" Mar 6 02:39:56.246146 containerd[1569]: 2026-03-06 02:39:56.122 [INFO][4705] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie188408e6a6 ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Namespace="kube-system" Pod="coredns-7d764666f9-rghxd" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--rghxd-eth0" Mar 6 02:39:56.246146 containerd[1569]: 2026-03-06 02:39:56.130 [INFO][4705] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Namespace="kube-system" Pod="coredns-7d764666f9-rghxd" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--rghxd-eth0" Mar 6 02:39:56.246146 containerd[1569]: 2026-03-06 02:39:56.131 [INFO][4705] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Namespace="kube-system" Pod="coredns-7d764666f9-rghxd" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--rghxd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--rghxd-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"48337397-07a0-44d5-b732-338087eee7ba", ResourceVersion:"1036", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 37, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f", Pod:"coredns-7d764666f9-rghxd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calie188408e6a6", MAC:"32:9a:0d:9c:96:5c", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:56.246146 containerd[1569]: 2026-03-06 02:39:56.233 [INFO][4705] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" Namespace="kube-system" Pod="coredns-7d764666f9-rghxd" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--rghxd-eth0" Mar 6 02:39:56.258082 containerd[1569]: time="2026-03-06T02:39:56.257988133Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5cf8bb7c94-m8khs,Uid:2776157f-9f5c-4064-8095-991904ea8002,Namespace:calico-system,Attempt:0,} returns sandbox id \"db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875\"" Mar 6 02:39:56.284139 containerd[1569]: time="2026-03-06T02:39:56.283877992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-r6hk4,Uid:ee80ff4d-1458-4649-bc23-11b25825a7c0,Namespace:kube-system,Attempt:0,}" Mar 6 02:39:56.368946 systemd-networkd[1497]: calid5572191cf2: Gained IPv6LL Mar 6 02:39:56.468697 containerd[1569]: time="2026-03-06T02:39:56.468447149Z" level=info msg="connecting to shim 2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f" address="unix:///run/containerd/s/64f175dc723eed0112bebc563dcace3b6680b16631d27fedc94171489c3b62b3" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:39:56.603721 systemd[1]: Started cri-containerd-2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f.scope - libcontainer container 2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f. Mar 6 02:39:56.682131 systemd-resolved[1399]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 02:39:56.847983 containerd[1569]: time="2026-03-06T02:39:56.847853186Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-rghxd,Uid:48337397-07a0-44d5-b732-338087eee7ba,Namespace:kube-system,Attempt:0,} returns sandbox id \"2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f\"" Mar 6 02:39:56.868886 containerd[1569]: time="2026-03-06T02:39:56.868169686Z" level=info msg="CreateContainer within sandbox \"2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 6 02:39:56.943491 kubelet[2879]: I0306 02:39:56.943067 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/whisker-57c85f46f-59gzc" podStartSLOduration=6.375456455 podStartE2EDuration="10.943049867s" podCreationTimestamp="2026-03-06 02:39:46 +0000 UTC" firstStartedPulling="2026-03-06 02:39:50.665940477 +0000 UTC m=+126.673138979" lastFinishedPulling="2026-03-06 02:39:55.233533899 +0000 UTC m=+131.240732391" observedRunningTime="2026-03-06 02:39:56.94169935 +0000 UTC m=+132.948897862" watchObservedRunningTime="2026-03-06 02:39:56.943049867 +0000 UTC m=+132.950248359" Mar 6 02:39:57.047198 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount402491469.mount: Deactivated successfully. Mar 6 02:39:57.138650 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1124310858.mount: Deactivated successfully. Mar 6 02:39:57.141901 containerd[1569]: time="2026-03-06T02:39:57.140091602Z" level=info msg="Container 1fc5bbe0e47489728e83cc697332e9fb1c7b3cbb4fb86eac5ff84a8a96631519: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:39:57.188562 containerd[1569]: time="2026-03-06T02:39:57.186049579Z" level=info msg="CreateContainer within sandbox \"2404fc6e294c7389db62ae68fcf42032f876a9d3a9f1098c06f79cf3de1ce16f\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"1fc5bbe0e47489728e83cc697332e9fb1c7b3cbb4fb86eac5ff84a8a96631519\"" Mar 6 02:39:57.192896 containerd[1569]: time="2026-03-06T02:39:57.191921741Z" level=info msg="StartContainer for \"1fc5bbe0e47489728e83cc697332e9fb1c7b3cbb4fb86eac5ff84a8a96631519\"" Mar 6 02:39:57.202886 containerd[1569]: time="2026-03-06T02:39:57.202066245Z" level=info msg="connecting to shim 1fc5bbe0e47489728e83cc697332e9fb1c7b3cbb4fb86eac5ff84a8a96631519" address="unix:///run/containerd/s/64f175dc723eed0112bebc563dcace3b6680b16631d27fedc94171489c3b62b3" protocol=ttrpc version=3 Mar 6 02:39:57.284170 systemd[1]: Started cri-containerd-1fc5bbe0e47489728e83cc697332e9fb1c7b3cbb4fb86eac5ff84a8a96631519.scope - libcontainer container 1fc5bbe0e47489728e83cc697332e9fb1c7b3cbb4fb86eac5ff84a8a96631519. Mar 6 02:39:57.362531 systemd-networkd[1497]: califf865893ab3: Link UP Mar 6 02:39:57.366466 systemd-networkd[1497]: califf865893ab3: Gained carrier Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.533 [INFO][4842] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7d764666f9--r6hk4-eth0 coredns-7d764666f9- kube-system ee80ff4d-1458-4649-bc23-11b25825a7c0 1044 0 2026-03-06 02:37:45 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7d764666f9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7d764666f9-r6hk4 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] califf865893ab3 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Namespace="kube-system" Pod="coredns-7d764666f9-r6hk4" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--r6hk4-" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.533 [INFO][4842] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Namespace="kube-system" Pod="coredns-7d764666f9-r6hk4" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.848 [INFO][4893] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" HandleID="k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Workload="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.899 [INFO][4893] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" HandleID="k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Workload="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003ecf20), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7d764666f9-r6hk4", "timestamp":"2026-03-06 02:39:56.848606872 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000366420)} Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.899 [INFO][4893] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.900 [INFO][4893] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.901 [INFO][4893] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.925 [INFO][4893] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:56.969 [INFO][4893] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.128 [INFO][4893] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.144 [INFO][4893] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.158 [INFO][4893] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.162 [INFO][4893] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.176 [INFO][4893] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.199 [INFO][4893] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.262 [INFO][4893] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.280 [INFO][4893] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" host="localhost" Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.280 [INFO][4893] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 6 02:39:57.492855 containerd[1569]: 2026-03-06 02:39:57.280 [INFO][4893] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" HandleID="k8s-pod-network.1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Workload="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" Mar 6 02:39:57.494490 containerd[1569]: 2026-03-06 02:39:57.291 [INFO][4842] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Namespace="kube-system" Pod="coredns-7d764666f9-r6hk4" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--r6hk4-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"ee80ff4d-1458-4649-bc23-11b25825a7c0", ResourceVersion:"1044", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 37, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7d764666f9-r6hk4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"califf865893ab3", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:57.494490 containerd[1569]: 2026-03-06 02:39:57.296 [INFO][4842] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Namespace="kube-system" Pod="coredns-7d764666f9-r6hk4" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" Mar 6 02:39:57.494490 containerd[1569]: 2026-03-06 02:39:57.296 [INFO][4842] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to califf865893ab3 ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Namespace="kube-system" Pod="coredns-7d764666f9-r6hk4" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" Mar 6 02:39:57.494490 containerd[1569]: 2026-03-06 02:39:57.366 [INFO][4842] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Namespace="kube-system" Pod="coredns-7d764666f9-r6hk4" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" Mar 6 02:39:57.494490 containerd[1569]: 2026-03-06 02:39:57.367 [INFO][4842] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Namespace="kube-system" Pod="coredns-7d764666f9-r6hk4" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--r6hk4-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"ee80ff4d-1458-4649-bc23-11b25825a7c0", ResourceVersion:"1044", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 37, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce", Pod:"coredns-7d764666f9-r6hk4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"califf865893ab3", MAC:"aa:50:a2:30:f3:92", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:57.494490 containerd[1569]: 2026-03-06 02:39:57.472 [INFO][4842] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" Namespace="kube-system" Pod="coredns-7d764666f9-r6hk4" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--r6hk4-eth0" Mar 6 02:39:57.646104 containerd[1569]: time="2026-03-06T02:39:57.645850186Z" level=info msg="connecting to shim 1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce" address="unix:///run/containerd/s/d342c2962b36f554b336091f9d34d62c40c407b47eacf08cdc191366570f50ee" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:39:57.729237 containerd[1569]: time="2026-03-06T02:39:57.728781037Z" level=info msg="StartContainer for \"1fc5bbe0e47489728e83cc697332e9fb1c7b3cbb4fb86eac5ff84a8a96631519\" returns successfully" Mar 6 02:39:57.763865 systemd[1]: Started cri-containerd-1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce.scope - libcontainer container 1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce. Mar 6 02:39:57.840068 systemd-resolved[1399]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 02:39:57.978473 kubelet[2879]: I0306 02:39:57.978087 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/coredns-7d764666f9-rghxd" podStartSLOduration=132.978063827 podStartE2EDuration="2m12.978063827s" podCreationTimestamp="2026-03-06 02:37:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 02:39:57.977102329 +0000 UTC m=+133.984300901" watchObservedRunningTime="2026-03-06 02:39:57.978063827 +0000 UTC m=+133.985262319" Mar 6 02:39:57.990657 containerd[1569]: time="2026-03-06T02:39:57.989082746Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-r6hk4,Uid:ee80ff4d-1458-4649-bc23-11b25825a7c0,Namespace:kube-system,Attempt:0,} returns sandbox id \"1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce\"" Mar 6 02:39:58.024957 containerd[1569]: time="2026-03-06T02:39:58.023668849Z" level=info msg="CreateContainer within sandbox \"1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 6 02:39:58.074443 containerd[1569]: time="2026-03-06T02:39:58.067621900Z" level=info msg="Container 015c612e81161052f533127ffc5f7b9b2e4452d50e4b7298ae340e25d92582f8: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:39:58.087066 containerd[1569]: time="2026-03-06T02:39:58.086965088Z" level=info msg="CreateContainer within sandbox \"1e0ed93e268af83aa9ba0b13bf9479e38fd7cdc7a7ad512d5fedf66365102bce\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"015c612e81161052f533127ffc5f7b9b2e4452d50e4b7298ae340e25d92582f8\"" Mar 6 02:39:58.091653 containerd[1569]: time="2026-03-06T02:39:58.090008145Z" level=info msg="StartContainer for \"015c612e81161052f533127ffc5f7b9b2e4452d50e4b7298ae340e25d92582f8\"" Mar 6 02:39:58.091653 containerd[1569]: time="2026-03-06T02:39:58.091193053Z" level=info msg="connecting to shim 015c612e81161052f533127ffc5f7b9b2e4452d50e4b7298ae340e25d92582f8" address="unix:///run/containerd/s/d342c2962b36f554b336091f9d34d62c40c407b47eacf08cdc191366570f50ee" protocol=ttrpc version=3 Mar 6 02:39:58.156592 systemd-networkd[1497]: calie188408e6a6: Gained IPv6LL Mar 6 02:39:58.199119 systemd[1]: Started cri-containerd-015c612e81161052f533127ffc5f7b9b2e4452d50e4b7298ae340e25d92582f8.scope - libcontainer container 015c612e81161052f533127ffc5f7b9b2e4452d50e4b7298ae340e25d92582f8. Mar 6 02:39:58.292639 containerd[1569]: time="2026-03-06T02:39:58.292220832Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-9f7667bb8-bz8j9,Uid:64f79500-120a-4333-adde-31efeaa36e77,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:58.298802 containerd[1569]: time="2026-03-06T02:39:58.293122904Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-d54f5f4cf-t2lk8,Uid:5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f,Namespace:calico-system,Attempt:0,}" Mar 6 02:39:58.435850 containerd[1569]: time="2026-03-06T02:39:58.435563015Z" level=info msg="StartContainer for \"015c612e81161052f533127ffc5f7b9b2e4452d50e4b7298ae340e25d92582f8\" returns successfully" Mar 6 02:39:58.979089 kubelet[2879]: I0306 02:39:58.978121 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/coredns-7d764666f9-r6hk4" podStartSLOduration=133.978103919 podStartE2EDuration="2m13.978103919s" podCreationTimestamp="2026-03-06 02:37:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 02:39:58.975885369 +0000 UTC m=+134.983083860" watchObservedRunningTime="2026-03-06 02:39:58.978103919 +0000 UTC m=+134.985302441" Mar 6 02:39:59.185552 systemd-networkd[1497]: caliad7e20aff45: Link UP Mar 6 02:39:59.185859 systemd-networkd[1497]: caliad7e20aff45: Gained carrier Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.625 [INFO][5072] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0 calico-apiserver-d54f5f4cf- calico-system 5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f 1042 0 2026-03-06 02:38:32 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:d54f5f4cf projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-d54f5f4cf-t2lk8 eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] caliad7e20aff45 [] [] }} ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-t2lk8" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.627 [INFO][5072] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-t2lk8" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.806 [INFO][5113] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" HandleID="k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Workload="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.837 [INFO][5113] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" HandleID="k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Workload="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000770de0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-apiserver-d54f5f4cf-t2lk8", "timestamp":"2026-03-06 02:39:58.806922517 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00026a2c0)} Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.837 [INFO][5113] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.837 [INFO][5113] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.837 [INFO][5113] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.857 [INFO][5113] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.882 [INFO][5113] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.938 [INFO][5113] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.965 [INFO][5113] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.991 [INFO][5113] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:58.991 [INFO][5113] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:59.023 [INFO][5113] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:59.068 [INFO][5113] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:59.131 [INFO][5113] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:59.134 [INFO][5113] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" host="localhost" Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:59.137 [INFO][5113] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 6 02:39:59.289749 containerd[1569]: 2026-03-06 02:39:59.143 [INFO][5113] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" HandleID="k8s-pod-network.f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Workload="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" Mar 6 02:39:59.292752 containerd[1569]: 2026-03-06 02:39:59.165 [INFO][5072] cni-plugin/k8s.go 418: Populated endpoint ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-t2lk8" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0", GenerateName:"calico-apiserver-d54f5f4cf-", Namespace:"calico-system", SelfLink:"", UID:"5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f", ResourceVersion:"1042", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"d54f5f4cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-d54f5f4cf-t2lk8", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"caliad7e20aff45", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:59.292752 containerd[1569]: 2026-03-06 02:39:59.167 [INFO][5072] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-t2lk8" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" Mar 6 02:39:59.292752 containerd[1569]: 2026-03-06 02:39:59.168 [INFO][5072] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliad7e20aff45 ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-t2lk8" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" Mar 6 02:39:59.292752 containerd[1569]: 2026-03-06 02:39:59.197 [INFO][5072] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-t2lk8" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" Mar 6 02:39:59.292752 containerd[1569]: 2026-03-06 02:39:59.199 [INFO][5072] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-t2lk8" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0", GenerateName:"calico-apiserver-d54f5f4cf-", Namespace:"calico-system", SelfLink:"", UID:"5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f", ResourceVersion:"1042", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"d54f5f4cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e", Pod:"calico-apiserver-d54f5f4cf-t2lk8", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"caliad7e20aff45", MAC:"8e:d2:a2:17:54:d9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:59.292752 containerd[1569]: 2026-03-06 02:39:59.260 [INFO][5072] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" Namespace="calico-system" Pod="calico-apiserver-d54f5f4cf-t2lk8" WorkloadEndpoint="localhost-k8s-calico--apiserver--d54f5f4cf--t2lk8-eth0" Mar 6 02:39:59.328042 systemd-networkd[1497]: califf865893ab3: Gained IPv6LL Mar 6 02:39:59.490452 containerd[1569]: time="2026-03-06T02:39:59.490154061Z" level=info msg="connecting to shim f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e" address="unix:///run/containerd/s/0f94728039ea53b7cba9c1cc2439eab204a430f94a09fa8c98d1b3144af0ea8b" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:39:59.638415 systemd[1]: Started cri-containerd-f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e.scope - libcontainer container f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e. Mar 6 02:39:59.675114 systemd-networkd[1497]: cali273962145a1: Link UP Mar 6 02:39:59.678397 systemd-networkd[1497]: cali273962145a1: Gained carrier Mar 6 02:39:59.782946 systemd-resolved[1399]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:58.756 [INFO][5064] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0 goldmane-9f7667bb8- calico-system 64f79500-120a-4333-adde-31efeaa36e77 1046 0 2026-03-06 02:38:33 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:9f7667bb8 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-9f7667bb8-bz8j9 eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali273962145a1 [] [] }} ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Namespace="calico-system" Pod="goldmane-9f7667bb8-bz8j9" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--bz8j9-" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:58.758 [INFO][5064] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Namespace="calico-system" Pod="goldmane-9f7667bb8-bz8j9" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:58.900 [INFO][5120] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" HandleID="k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Workload="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:58.931 [INFO][5120] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" HandleID="k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Workload="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000478120), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-9f7667bb8-bz8j9", "timestamp":"2026-03-06 02:39:58.900015666 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0003a1a20)} Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:58.932 [INFO][5120] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.136 [INFO][5120] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.144 [INFO][5120] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.267 [INFO][5120] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.311 [INFO][5120] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.381 [INFO][5120] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.395 [INFO][5120] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.428 [INFO][5120] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.429 [INFO][5120] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.443 [INFO][5120] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.539 [INFO][5120] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.620 [INFO][5120] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.620 [INFO][5120] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" host="localhost" Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.620 [INFO][5120] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 6 02:39:59.833255 containerd[1569]: 2026-03-06 02:39:59.620 [INFO][5120] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" HandleID="k8s-pod-network.0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Workload="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" Mar 6 02:39:59.835608 containerd[1569]: 2026-03-06 02:39:59.639 [INFO][5064] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Namespace="calico-system" Pod="goldmane-9f7667bb8-bz8j9" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0", GenerateName:"goldmane-9f7667bb8-", Namespace:"calico-system", SelfLink:"", UID:"64f79500-120a-4333-adde-31efeaa36e77", ResourceVersion:"1046", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"9f7667bb8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-9f7667bb8-bz8j9", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali273962145a1", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:59.835608 containerd[1569]: 2026-03-06 02:39:59.640 [INFO][5064] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Namespace="calico-system" Pod="goldmane-9f7667bb8-bz8j9" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" Mar 6 02:39:59.835608 containerd[1569]: 2026-03-06 02:39:59.640 [INFO][5064] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali273962145a1 ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Namespace="calico-system" Pod="goldmane-9f7667bb8-bz8j9" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" Mar 6 02:39:59.835608 containerd[1569]: 2026-03-06 02:39:59.682 [INFO][5064] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Namespace="calico-system" Pod="goldmane-9f7667bb8-bz8j9" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" Mar 6 02:39:59.835608 containerd[1569]: 2026-03-06 02:39:59.683 [INFO][5064] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Namespace="calico-system" Pod="goldmane-9f7667bb8-bz8j9" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0", GenerateName:"goldmane-9f7667bb8-", Namespace:"calico-system", SelfLink:"", UID:"64f79500-120a-4333-adde-31efeaa36e77", ResourceVersion:"1046", Generation:0, CreationTimestamp:time.Date(2026, time.March, 6, 2, 38, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"9f7667bb8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea", Pod:"goldmane-9f7667bb8-bz8j9", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali273962145a1", MAC:"c6:9b:03:22:e0:55", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 6 02:39:59.835608 containerd[1569]: 2026-03-06 02:39:59.800 [INFO][5064] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" Namespace="calico-system" Pod="goldmane-9f7667bb8-bz8j9" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--bz8j9-eth0" Mar 6 02:39:59.966605 containerd[1569]: time="2026-03-06T02:39:59.965147200Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-d54f5f4cf-t2lk8,Uid:5b1f7233-0ca1-4dd4-8a21-8dd7eac72e6f,Namespace:calico-system,Attempt:0,} returns sandbox id \"f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e\"" Mar 6 02:39:59.981891 containerd[1569]: time="2026-03-06T02:39:59.981631006Z" level=info msg="connecting to shim 0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea" address="unix:///run/containerd/s/9376226e6ee7a31554ed890772df18cabb4922a6aeb14fcb3976ff8aa67f207e" namespace=k8s.io protocol=ttrpc version=3 Mar 6 02:40:00.097654 systemd[1]: Started cri-containerd-0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea.scope - libcontainer container 0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea. Mar 6 02:40:00.161602 systemd[1]: Started sshd@7-10.0.0.69:22-10.0.0.1:41454.service - OpenSSH per-connection server daemon (10.0.0.1:41454). Mar 6 02:40:00.193830 systemd-resolved[1399]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 02:40:00.567061 containerd[1569]: time="2026-03-06T02:40:00.566847968Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-9f7667bb8-bz8j9,Uid:64f79500-120a-4333-adde-31efeaa36e77,Namespace:calico-system,Attempt:0,} returns sandbox id \"0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea\"" Mar 6 02:40:00.590445 sshd[5262]: Accepted publickey for core from 10.0.0.1 port 41454 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:00.616532 sshd-session[5262]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:00.656465 systemd-logind[1544]: New session 8 of user core. Mar 6 02:40:00.664631 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 6 02:40:01.036970 systemd-networkd[1497]: caliad7e20aff45: Gained IPv6LL Mar 6 02:40:01.172059 sshd[5278]: Connection closed by 10.0.0.1 port 41454 Mar 6 02:40:01.173651 sshd-session[5262]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:01.187201 systemd[1]: sshd@7-10.0.0.69:22-10.0.0.1:41454.service: Deactivated successfully. Mar 6 02:40:01.191585 systemd[1]: session-8.scope: Deactivated successfully. Mar 6 02:40:01.196723 systemd-logind[1544]: Session 8 logged out. Waiting for processes to exit. Mar 6 02:40:01.206903 systemd-logind[1544]: Removed session 8. Mar 6 02:40:01.353855 systemd-networkd[1497]: cali273962145a1: Gained IPv6LL Mar 6 02:40:02.200201 containerd[1569]: time="2026-03-06T02:40:02.199644389Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:02.206988 containerd[1569]: time="2026-03-06T02:40:02.206747126Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=48415780" Mar 6 02:40:02.229600 containerd[1569]: time="2026-03-06T02:40:02.229542902Z" level=info msg="ImageCreate event name:\"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:02.235040 containerd[1569]: time="2026-03-06T02:40:02.234856274Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:02.237466 containerd[1569]: time="2026-03-06T02:40:02.237264072Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"49971841\" in 6.977049783s" Mar 6 02:40:02.237466 containerd[1569]: time="2026-03-06T02:40:02.237456211Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\"" Mar 6 02:40:02.242246 containerd[1569]: time="2026-03-06T02:40:02.242053260Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\"" Mar 6 02:40:02.250431 containerd[1569]: time="2026-03-06T02:40:02.250212665Z" level=info msg="CreateContainer within sandbox \"4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 6 02:40:02.275868 containerd[1569]: time="2026-03-06T02:40:02.275069761Z" level=info msg="Container 73ab41b3c9bd35524ea5d80781a633d35e34eff9fb0ab707f4f21f4574d1cf12: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:40:02.283916 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1459161000.mount: Deactivated successfully. Mar 6 02:40:02.330490 containerd[1569]: time="2026-03-06T02:40:02.330106473Z" level=info msg="CreateContainer within sandbox \"4ab2de1cd6511cc55bb4e1a6fea94dc59830c62c9fc448399826aa25421c128d\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"73ab41b3c9bd35524ea5d80781a633d35e34eff9fb0ab707f4f21f4574d1cf12\"" Mar 6 02:40:02.332852 containerd[1569]: time="2026-03-06T02:40:02.332071180Z" level=info msg="StartContainer for \"73ab41b3c9bd35524ea5d80781a633d35e34eff9fb0ab707f4f21f4574d1cf12\"" Mar 6 02:40:02.336609 containerd[1569]: time="2026-03-06T02:40:02.336569497Z" level=info msg="connecting to shim 73ab41b3c9bd35524ea5d80781a633d35e34eff9fb0ab707f4f21f4574d1cf12" address="unix:///run/containerd/s/808e894657a328f829b0e8819ca1522ddcf88fe34e7a257dbb4241b470dae5b1" protocol=ttrpc version=3 Mar 6 02:40:02.411001 systemd[1]: Started cri-containerd-73ab41b3c9bd35524ea5d80781a633d35e34eff9fb0ab707f4f21f4574d1cf12.scope - libcontainer container 73ab41b3c9bd35524ea5d80781a633d35e34eff9fb0ab707f4f21f4574d1cf12. Mar 6 02:40:02.585229 containerd[1569]: time="2026-03-06T02:40:02.585003751Z" level=info msg="StartContainer for \"73ab41b3c9bd35524ea5d80781a633d35e34eff9fb0ab707f4f21f4574d1cf12\" returns successfully" Mar 6 02:40:04.344245 containerd[1569]: time="2026-03-06T02:40:04.343979213Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:04.348722 containerd[1569]: time="2026-03-06T02:40:04.348625145Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.4: active requests=0, bytes read=8792502" Mar 6 02:40:04.354035 containerd[1569]: time="2026-03-06T02:40:04.352966399Z" level=info msg="ImageCreate event name:\"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:04.361905 containerd[1569]: time="2026-03-06T02:40:04.361800961Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.4\" with image id \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\", size \"10348547\" in 2.119708398s" Mar 6 02:40:04.362018 containerd[1569]: time="2026-03-06T02:40:04.361897852Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\" returns image reference \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\"" Mar 6 02:40:04.362150 containerd[1569]: time="2026-03-06T02:40:04.362115316Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:04.366428 containerd[1569]: time="2026-03-06T02:40:04.366191796Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\"" Mar 6 02:40:04.374889 containerd[1569]: time="2026-03-06T02:40:04.374563506Z" level=info msg="CreateContainer within sandbox \"ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 6 02:40:04.420436 containerd[1569]: time="2026-03-06T02:40:04.418021769Z" level=info msg="Container a53cceb8de723c7994b9ff80f7264654ca077e504f908378a148210701f93dfb: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:40:04.447968 containerd[1569]: time="2026-03-06T02:40:04.447857507Z" level=info msg="CreateContainer within sandbox \"ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"a53cceb8de723c7994b9ff80f7264654ca077e504f908378a148210701f93dfb\"" Mar 6 02:40:04.449752 containerd[1569]: time="2026-03-06T02:40:04.449520060Z" level=info msg="StartContainer for \"a53cceb8de723c7994b9ff80f7264654ca077e504f908378a148210701f93dfb\"" Mar 6 02:40:04.452695 containerd[1569]: time="2026-03-06T02:40:04.452543458Z" level=info msg="connecting to shim a53cceb8de723c7994b9ff80f7264654ca077e504f908378a148210701f93dfb" address="unix:///run/containerd/s/704dd4bf980acf83e2286c33e0dfa7a716b21035f8f6a97e47896491667cd04d" protocol=ttrpc version=3 Mar 6 02:40:04.521165 systemd[1]: Started cri-containerd-a53cceb8de723c7994b9ff80f7264654ca077e504f908378a148210701f93dfb.scope - libcontainer container a53cceb8de723c7994b9ff80f7264654ca077e504f908378a148210701f93dfb. Mar 6 02:40:04.752872 containerd[1569]: time="2026-03-06T02:40:04.752538820Z" level=info msg="StartContainer for \"a53cceb8de723c7994b9ff80f7264654ca077e504f908378a148210701f93dfb\" returns successfully" Mar 6 02:40:04.993473 kubelet[2879]: I0306 02:40:04.993088 2879 prober_manager.go:356] "Failed to trigger a manual run" probe="Readiness" Mar 6 02:40:06.009022 kubelet[2879]: I0306 02:40:06.008167 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-apiserver-d54f5f4cf-vrw2m" podStartSLOduration=86.509705738 podStartE2EDuration="1m34.008145219s" podCreationTimestamp="2026-03-06 02:38:32 +0000 UTC" firstStartedPulling="2026-03-06 02:39:54.741685011 +0000 UTC m=+130.748883503" lastFinishedPulling="2026-03-06 02:40:02.240124491 +0000 UTC m=+138.247322984" observedRunningTime="2026-03-06 02:40:02.990688539 +0000 UTC m=+138.997887031" watchObservedRunningTime="2026-03-06 02:40:06.008145219 +0000 UTC m=+142.015343711" Mar 6 02:40:06.194417 systemd[1]: Started sshd@8-10.0.0.69:22-10.0.0.1:33570.service - OpenSSH per-connection server daemon (10.0.0.1:33570). Mar 6 02:40:06.652804 sshd[5406]: Accepted publickey for core from 10.0.0.1 port 33570 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:06.658937 sshd-session[5406]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:06.680254 systemd-logind[1544]: New session 9 of user core. Mar 6 02:40:06.691213 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 6 02:40:07.038816 sshd[5413]: Connection closed by 10.0.0.1 port 33570 Mar 6 02:40:07.038466 sshd-session[5406]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:07.045814 systemd[1]: sshd@8-10.0.0.69:22-10.0.0.1:33570.service: Deactivated successfully. Mar 6 02:40:07.052804 systemd[1]: session-9.scope: Deactivated successfully. Mar 6 02:40:07.054979 systemd-logind[1544]: Session 9 logged out. Waiting for processes to exit. Mar 6 02:40:07.061815 systemd-logind[1544]: Removed session 9. Mar 6 02:40:10.945995 containerd[1569]: time="2026-03-06T02:40:10.945621584Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:10.949074 containerd[1569]: time="2026-03-06T02:40:10.949030198Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.4: active requests=0, bytes read=52406348" Mar 6 02:40:10.957551 containerd[1569]: time="2026-03-06T02:40:10.956636218Z" level=info msg="ImageCreate event name:\"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:10.967488 containerd[1569]: time="2026-03-06T02:40:10.967228739Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:10.971765 containerd[1569]: time="2026-03-06T02:40:10.971634751Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" with image id \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\", size \"53962361\" in 6.605396358s" Mar 6 02:40:10.971899 containerd[1569]: time="2026-03-06T02:40:10.971775565Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" returns image reference \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\"" Mar 6 02:40:10.983565 containerd[1569]: time="2026-03-06T02:40:10.983124460Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Mar 6 02:40:11.043741 containerd[1569]: time="2026-03-06T02:40:11.043685084Z" level=info msg="CreateContainer within sandbox \"db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Mar 6 02:40:11.089255 containerd[1569]: time="2026-03-06T02:40:11.089207085Z" level=info msg="Container 1684ef6876093390b1c5dc26523275cdf322e199a2b531b241f28a97498a6ae1: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:40:11.125169 containerd[1569]: time="2026-03-06T02:40:11.124805989Z" level=info msg="CreateContainer within sandbox \"db46853d51cdf3bc6a7f00e29315bb538294e160d2ef6418546e6cb40433a875\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"1684ef6876093390b1c5dc26523275cdf322e199a2b531b241f28a97498a6ae1\"" Mar 6 02:40:11.129580 containerd[1569]: time="2026-03-06T02:40:11.126922775Z" level=info msg="StartContainer for \"1684ef6876093390b1c5dc26523275cdf322e199a2b531b241f28a97498a6ae1\"" Mar 6 02:40:11.133627 containerd[1569]: time="2026-03-06T02:40:11.133588134Z" level=info msg="connecting to shim 1684ef6876093390b1c5dc26523275cdf322e199a2b531b241f28a97498a6ae1" address="unix:///run/containerd/s/df6d070569a4d3df533c26aa7e25933ab104d88917e7e55ee0eee91410ea798a" protocol=ttrpc version=3 Mar 6 02:40:11.236204 systemd[1]: Started cri-containerd-1684ef6876093390b1c5dc26523275cdf322e199a2b531b241f28a97498a6ae1.scope - libcontainer container 1684ef6876093390b1c5dc26523275cdf322e199a2b531b241f28a97498a6ae1. Mar 6 02:40:11.254236 containerd[1569]: time="2026-03-06T02:40:11.254192208Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:11.257922 containerd[1569]: time="2026-03-06T02:40:11.257890878Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=77" Mar 6 02:40:11.286486 containerd[1569]: time="2026-03-06T02:40:11.285634003Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"49971841\" in 296.040229ms" Mar 6 02:40:11.286486 containerd[1569]: time="2026-03-06T02:40:11.285694416Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\"" Mar 6 02:40:11.291673 containerd[1569]: time="2026-03-06T02:40:11.291508644Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\"" Mar 6 02:40:11.325193 containerd[1569]: time="2026-03-06T02:40:11.324707102Z" level=info msg="CreateContainer within sandbox \"f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 6 02:40:11.364779 containerd[1569]: time="2026-03-06T02:40:11.364648393Z" level=info msg="Container 28a041dec6d311a4f96002dd614c0e358e0f3f4aaa83623527d2770f6b1c05d2: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:40:11.402587 containerd[1569]: time="2026-03-06T02:40:11.402230401Z" level=info msg="CreateContainer within sandbox \"f9f7cbd0d1ac68f170534d304fe9098d7219eeb998a248b998ae13b9eb1a966e\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"28a041dec6d311a4f96002dd614c0e358e0f3f4aaa83623527d2770f6b1c05d2\"" Mar 6 02:40:11.408951 containerd[1569]: time="2026-03-06T02:40:11.408031909Z" level=info msg="StartContainer for \"28a041dec6d311a4f96002dd614c0e358e0f3f4aaa83623527d2770f6b1c05d2\"" Mar 6 02:40:11.424601 containerd[1569]: time="2026-03-06T02:40:11.424261470Z" level=info msg="connecting to shim 28a041dec6d311a4f96002dd614c0e358e0f3f4aaa83623527d2770f6b1c05d2" address="unix:///run/containerd/s/0f94728039ea53b7cba9c1cc2439eab204a430f94a09fa8c98d1b3144af0ea8b" protocol=ttrpc version=3 Mar 6 02:40:11.528071 containerd[1569]: time="2026-03-06T02:40:11.527687304Z" level=info msg="StartContainer for \"1684ef6876093390b1c5dc26523275cdf322e199a2b531b241f28a97498a6ae1\" returns successfully" Mar 6 02:40:11.543198 systemd[1]: Started cri-containerd-28a041dec6d311a4f96002dd614c0e358e0f3f4aaa83623527d2770f6b1c05d2.scope - libcontainer container 28a041dec6d311a4f96002dd614c0e358e0f3f4aaa83623527d2770f6b1c05d2. Mar 6 02:40:11.883153 containerd[1569]: time="2026-03-06T02:40:11.883072376Z" level=info msg="StartContainer for \"28a041dec6d311a4f96002dd614c0e358e0f3f4aaa83623527d2770f6b1c05d2\" returns successfully" Mar 6 02:40:12.071843 systemd[1]: Started sshd@9-10.0.0.69:22-10.0.0.1:42696.service - OpenSSH per-connection server daemon (10.0.0.1:42696). Mar 6 02:40:12.262009 kubelet[2879]: I0306 02:40:12.261736 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-apiserver-d54f5f4cf-t2lk8" podStartSLOduration=88.947529418 podStartE2EDuration="1m40.26171396s" podCreationTimestamp="2026-03-06 02:38:32 +0000 UTC" firstStartedPulling="2026-03-06 02:39:59.975670489 +0000 UTC m=+135.982869011" lastFinishedPulling="2026-03-06 02:40:11.289855061 +0000 UTC m=+147.297053553" observedRunningTime="2026-03-06 02:40:12.25280358 +0000 UTC m=+148.260002072" watchObservedRunningTime="2026-03-06 02:40:12.26171396 +0000 UTC m=+148.268912472" Mar 6 02:40:12.407589 sshd[5510]: Accepted publickey for core from 10.0.0.1 port 42696 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:12.410168 sshd-session[5510]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:12.437650 systemd-logind[1544]: New session 10 of user core. Mar 6 02:40:12.454249 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 6 02:40:12.617151 kubelet[2879]: I0306 02:40:12.616953 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5cf8bb7c94-m8khs" podStartSLOduration=82.92149939 podStartE2EDuration="1m37.616934029s" podCreationTimestamp="2026-03-06 02:38:35 +0000 UTC" firstStartedPulling="2026-03-06 02:39:56.279754154 +0000 UTC m=+132.286952656" lastFinishedPulling="2026-03-06 02:40:10.975188793 +0000 UTC m=+146.982387295" observedRunningTime="2026-03-06 02:40:12.361276327 +0000 UTC m=+148.368474829" watchObservedRunningTime="2026-03-06 02:40:12.616934029 +0000 UTC m=+148.624132522" Mar 6 02:40:13.184575 sshd[5544]: Connection closed by 10.0.0.1 port 42696 Mar 6 02:40:13.185844 sshd-session[5510]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:13.195719 systemd[1]: sshd@9-10.0.0.69:22-10.0.0.1:42696.service: Deactivated successfully. Mar 6 02:40:13.197777 systemd-logind[1544]: Session 10 logged out. Waiting for processes to exit. Mar 6 02:40:13.202221 systemd[1]: session-10.scope: Deactivated successfully. Mar 6 02:40:13.227587 systemd-logind[1544]: Removed session 10. Mar 6 02:40:16.694966 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4037215209.mount: Deactivated successfully. Mar 6 02:40:18.215004 systemd[1]: Started sshd@10-10.0.0.69:22-10.0.0.1:42698.service - OpenSSH per-connection server daemon (10.0.0.1:42698). Mar 6 02:40:18.546506 sshd[5611]: Accepted publickey for core from 10.0.0.1 port 42698 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:18.556051 sshd-session[5611]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:18.579193 systemd-logind[1544]: New session 11 of user core. Mar 6 02:40:18.598093 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 6 02:40:19.800907 containerd[1569]: time="2026-03-06T02:40:19.800854164Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:19.815035 containerd[1569]: time="2026-03-06T02:40:19.803066942Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.4: active requests=0, bytes read=55623386" Mar 6 02:40:19.815035 containerd[1569]: time="2026-03-06T02:40:19.806214372Z" level=info msg="ImageCreate event name:\"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:19.818657 containerd[1569]: time="2026-03-06T02:40:19.818616280Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:19.822201 containerd[1569]: time="2026-03-06T02:40:19.822164427Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" with image id \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\", size \"55623232\" in 8.530553703s" Mar 6 02:40:19.823046 containerd[1569]: time="2026-03-06T02:40:19.823020007Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" returns image reference \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\"" Mar 6 02:40:19.846699 containerd[1569]: time="2026-03-06T02:40:19.845605474Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\"" Mar 6 02:40:19.880872 sshd[5614]: Connection closed by 10.0.0.1 port 42698 Mar 6 02:40:19.884064 sshd-session[5611]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:19.889458 containerd[1569]: time="2026-03-06T02:40:19.887231506Z" level=info msg="CreateContainer within sandbox \"0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Mar 6 02:40:19.902834 systemd[1]: sshd@10-10.0.0.69:22-10.0.0.1:42698.service: Deactivated successfully. Mar 6 02:40:19.920972 systemd[1]: session-11.scope: Deactivated successfully. Mar 6 02:40:19.925174 systemd-logind[1544]: Session 11 logged out. Waiting for processes to exit. Mar 6 02:40:19.929173 systemd-logind[1544]: Removed session 11. Mar 6 02:40:19.974527 containerd[1569]: time="2026-03-06T02:40:19.971597288Z" level=info msg="Container 71d3a8be6e4db083ec0bd9e98c0872faf0a9f7c00436f4e221411a512a7d7b2a: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:40:20.050726 containerd[1569]: time="2026-03-06T02:40:20.049942881Z" level=info msg="CreateContainer within sandbox \"0d17a9a2ca81a3b8e9a8fcf0026b9e60ca094261834b3d4caa61ac71815f83ea\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"71d3a8be6e4db083ec0bd9e98c0872faf0a9f7c00436f4e221411a512a7d7b2a\"" Mar 6 02:40:20.054635 containerd[1569]: time="2026-03-06T02:40:20.052774958Z" level=info msg="StartContainer for \"71d3a8be6e4db083ec0bd9e98c0872faf0a9f7c00436f4e221411a512a7d7b2a\"" Mar 6 02:40:20.066857 containerd[1569]: time="2026-03-06T02:40:20.066670413Z" level=info msg="connecting to shim 71d3a8be6e4db083ec0bd9e98c0872faf0a9f7c00436f4e221411a512a7d7b2a" address="unix:///run/containerd/s/9376226e6ee7a31554ed890772df18cabb4922a6aeb14fcb3976ff8aa67f207e" protocol=ttrpc version=3 Mar 6 02:40:20.185741 systemd[1]: Started cri-containerd-71d3a8be6e4db083ec0bd9e98c0872faf0a9f7c00436f4e221411a512a7d7b2a.scope - libcontainer container 71d3a8be6e4db083ec0bd9e98c0872faf0a9f7c00436f4e221411a512a7d7b2a. Mar 6 02:40:20.537723 containerd[1569]: time="2026-03-06T02:40:20.537196305Z" level=info msg="StartContainer for \"71d3a8be6e4db083ec0bd9e98c0872faf0a9f7c00436f4e221411a512a7d7b2a\" returns successfully" Mar 6 02:40:22.936012 containerd[1569]: time="2026-03-06T02:40:22.935243863Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:22.938704 containerd[1569]: time="2026-03-06T02:40:22.938575810Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4: active requests=0, bytes read=14704317" Mar 6 02:40:22.942100 containerd[1569]: time="2026-03-06T02:40:22.941941434Z" level=info msg="ImageCreate event name:\"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:22.948182 containerd[1569]: time="2026-03-06T02:40:22.947851874Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 02:40:22.950118 containerd[1569]: time="2026-03-06T02:40:22.950008273Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" with image id \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\", size \"16260314\" in 3.102895661s" Mar 6 02:40:22.950118 containerd[1569]: time="2026-03-06T02:40:22.950049581Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" returns image reference \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\"" Mar 6 02:40:22.967968 containerd[1569]: time="2026-03-06T02:40:22.967599788Z" level=info msg="CreateContainer within sandbox \"ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 6 02:40:22.994892 containerd[1569]: time="2026-03-06T02:40:22.994522496Z" level=info msg="Container 4cb62c45a2f8162d85f7d3390db4411726042b4c5def571ba65190448f8a0f6b: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:40:23.063450 containerd[1569]: time="2026-03-06T02:40:23.063174394Z" level=info msg="CreateContainer within sandbox \"ad5726ae7d703e013699384146139c0a5d35bcd3bbfe441cbd39b320a101d889\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"4cb62c45a2f8162d85f7d3390db4411726042b4c5def571ba65190448f8a0f6b\"" Mar 6 02:40:23.069962 containerd[1569]: time="2026-03-06T02:40:23.067973987Z" level=info msg="StartContainer for \"4cb62c45a2f8162d85f7d3390db4411726042b4c5def571ba65190448f8a0f6b\"" Mar 6 02:40:23.077207 containerd[1569]: time="2026-03-06T02:40:23.077103086Z" level=info msg="connecting to shim 4cb62c45a2f8162d85f7d3390db4411726042b4c5def571ba65190448f8a0f6b" address="unix:///run/containerd/s/704dd4bf980acf83e2286c33e0dfa7a716b21035f8f6a97e47896491667cd04d" protocol=ttrpc version=3 Mar 6 02:40:23.175878 systemd[1]: Started cri-containerd-4cb62c45a2f8162d85f7d3390db4411726042b4c5def571ba65190448f8a0f6b.scope - libcontainer container 4cb62c45a2f8162d85f7d3390db4411726042b4c5def571ba65190448f8a0f6b. Mar 6 02:40:23.388691 containerd[1569]: time="2026-03-06T02:40:23.388597057Z" level=info msg="StartContainer for \"4cb62c45a2f8162d85f7d3390db4411726042b4c5def571ba65190448f8a0f6b\" returns successfully" Mar 6 02:40:24.417664 kubelet[2879]: I0306 02:40:24.417083 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/csi-node-driver-gqfqd" podStartSLOduration=82.28925783 podStartE2EDuration="1m49.417068727s" podCreationTimestamp="2026-03-06 02:38:35 +0000 UTC" firstStartedPulling="2026-03-06 02:39:55.824882822 +0000 UTC m=+131.832081324" lastFinishedPulling="2026-03-06 02:40:22.952693729 +0000 UTC m=+158.959892221" observedRunningTime="2026-03-06 02:40:24.402952484 +0000 UTC m=+160.410151037" watchObservedRunningTime="2026-03-06 02:40:24.417068727 +0000 UTC m=+160.424267219" Mar 6 02:40:24.428694 kubelet[2879]: I0306 02:40:24.419891 2879 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/goldmane-9f7667bb8-bz8j9" podStartSLOduration=92.157648069 podStartE2EDuration="1m51.417719544s" podCreationTimestamp="2026-03-06 02:38:33 +0000 UTC" firstStartedPulling="2026-03-06 02:40:00.582923298 +0000 UTC m=+136.590121800" lastFinishedPulling="2026-03-06 02:40:19.842994783 +0000 UTC m=+155.850193275" observedRunningTime="2026-03-06 02:40:21.418844873 +0000 UTC m=+157.426043365" watchObservedRunningTime="2026-03-06 02:40:24.417719544 +0000 UTC m=+160.424918037" Mar 6 02:40:24.437148 kubelet[2879]: I0306 02:40:24.436994 2879 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 6 02:40:24.437148 kubelet[2879]: I0306 02:40:24.437051 2879 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 6 02:40:24.922005 systemd[1]: Started sshd@11-10.0.0.69:22-10.0.0.1:38898.service - OpenSSH per-connection server daemon (10.0.0.1:38898). Mar 6 02:40:25.282111 sshd[5788]: Accepted publickey for core from 10.0.0.1 port 38898 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:25.297587 sshd-session[5788]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:25.336692 systemd-logind[1544]: New session 12 of user core. Mar 6 02:40:25.349736 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 6 02:40:25.897256 sshd[5794]: Connection closed by 10.0.0.1 port 38898 Mar 6 02:40:25.897798 sshd-session[5788]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:25.909705 systemd[1]: sshd@11-10.0.0.69:22-10.0.0.1:38898.service: Deactivated successfully. Mar 6 02:40:25.916168 systemd[1]: session-12.scope: Deactivated successfully. Mar 6 02:40:25.921551 systemd-logind[1544]: Session 12 logged out. Waiting for processes to exit. Mar 6 02:40:25.925143 systemd-logind[1544]: Removed session 12. Mar 6 02:40:30.936612 systemd[1]: Started sshd@12-10.0.0.69:22-10.0.0.1:38914.service - OpenSSH per-connection server daemon (10.0.0.1:38914). Mar 6 02:40:31.068714 sshd[5811]: Accepted publickey for core from 10.0.0.1 port 38914 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:31.076875 sshd-session[5811]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:31.098824 systemd-logind[1544]: New session 13 of user core. Mar 6 02:40:31.119108 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 6 02:40:32.213457 sshd[5814]: Connection closed by 10.0.0.1 port 38914 Mar 6 02:40:32.654990 sshd-session[5811]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:33.248928 systemd[1]: sshd@12-10.0.0.69:22-10.0.0.1:38914.service: Deactivated successfully. Mar 6 02:40:33.273676 systemd[1]: session-13.scope: Deactivated successfully. Mar 6 02:40:33.289789 systemd-logind[1544]: Session 13 logged out. Waiting for processes to exit. Mar 6 02:40:33.298947 systemd-logind[1544]: Removed session 13. Mar 6 02:40:37.090583 systemd[1]: Started sshd@13-10.0.0.69:22-10.0.0.1:37998.service - OpenSSH per-connection server daemon (10.0.0.1:37998). Mar 6 02:40:37.369147 sshd[5843]: Accepted publickey for core from 10.0.0.1 port 37998 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:37.379226 sshd-session[5843]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:37.434658 systemd-logind[1544]: New session 14 of user core. Mar 6 02:40:37.440940 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 6 02:40:37.975169 sshd[5846]: Connection closed by 10.0.0.1 port 37998 Mar 6 02:40:37.979872 sshd-session[5843]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:37.993999 systemd[1]: sshd@13-10.0.0.69:22-10.0.0.1:37998.service: Deactivated successfully. Mar 6 02:40:38.002043 systemd[1]: session-14.scope: Deactivated successfully. Mar 6 02:40:38.008926 systemd-logind[1544]: Session 14 logged out. Waiting for processes to exit. Mar 6 02:40:38.023175 systemd-logind[1544]: Removed session 14. Mar 6 02:40:43.013781 systemd[1]: Started sshd@14-10.0.0.69:22-10.0.0.1:58436.service - OpenSSH per-connection server daemon (10.0.0.1:58436). Mar 6 02:40:43.219043 sshd[5939]: Accepted publickey for core from 10.0.0.1 port 58436 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:43.226164 sshd-session[5939]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:43.257047 systemd-logind[1544]: New session 15 of user core. Mar 6 02:40:43.274801 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 6 02:40:43.936737 sshd[5942]: Connection closed by 10.0.0.1 port 58436 Mar 6 02:40:43.936968 sshd-session[5939]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:43.947604 systemd[1]: sshd@14-10.0.0.69:22-10.0.0.1:58436.service: Deactivated successfully. Mar 6 02:40:43.955774 systemd[1]: session-15.scope: Deactivated successfully. Mar 6 02:40:43.964784 systemd-logind[1544]: Session 15 logged out. Waiting for processes to exit. Mar 6 02:40:43.972032 systemd-logind[1544]: Removed session 15. Mar 6 02:40:48.978827 systemd[1]: Started sshd@15-10.0.0.69:22-10.0.0.1:58444.service - OpenSSH per-connection server daemon (10.0.0.1:58444). Mar 6 02:40:49.164222 sshd[5983]: Accepted publickey for core from 10.0.0.1 port 58444 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:49.167105 sshd-session[5983]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:49.196386 systemd-logind[1544]: New session 16 of user core. Mar 6 02:40:49.231710 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 6 02:40:49.673058 sshd[5986]: Connection closed by 10.0.0.1 port 58444 Mar 6 02:40:49.674699 sshd-session[5983]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:49.698093 systemd[1]: sshd@15-10.0.0.69:22-10.0.0.1:58444.service: Deactivated successfully. Mar 6 02:40:49.705009 systemd[1]: session-16.scope: Deactivated successfully. Mar 6 02:40:49.716809 systemd-logind[1544]: Session 16 logged out. Waiting for processes to exit. Mar 6 02:40:49.726117 systemd-logind[1544]: Removed session 16. Mar 6 02:40:50.274951 kubelet[2879]: E0306 02:40:50.271866 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:40:54.699677 systemd[1]: Started sshd@16-10.0.0.69:22-10.0.0.1:39346.service - OpenSSH per-connection server daemon (10.0.0.1:39346). Mar 6 02:40:54.851071 sshd[6035]: Accepted publickey for core from 10.0.0.1 port 39346 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:40:54.855098 sshd-session[6035]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:40:54.870797 systemd-logind[1544]: New session 17 of user core. Mar 6 02:40:54.878807 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 6 02:40:55.182599 sshd[6038]: Connection closed by 10.0.0.1 port 39346 Mar 6 02:40:55.185239 sshd-session[6035]: pam_unix(sshd:session): session closed for user core Mar 6 02:40:55.200961 systemd[1]: sshd@16-10.0.0.69:22-10.0.0.1:39346.service: Deactivated successfully. Mar 6 02:40:55.210226 systemd[1]: session-17.scope: Deactivated successfully. Mar 6 02:40:55.213938 systemd-logind[1544]: Session 17 logged out. Waiting for processes to exit. Mar 6 02:40:55.219882 systemd-logind[1544]: Removed session 17. Mar 6 02:41:00.214214 systemd[1]: Started sshd@17-10.0.0.69:22-10.0.0.1:39362.service - OpenSSH per-connection server daemon (10.0.0.1:39362). Mar 6 02:41:00.340554 sshd[6055]: Accepted publickey for core from 10.0.0.1 port 39362 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:00.345212 sshd-session[6055]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:00.378641 systemd-logind[1544]: New session 18 of user core. Mar 6 02:41:00.397693 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 6 02:41:00.747129 sshd[6058]: Connection closed by 10.0.0.1 port 39362 Mar 6 02:41:00.752155 sshd-session[6055]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:00.770924 systemd[1]: sshd@17-10.0.0.69:22-10.0.0.1:39362.service: Deactivated successfully. Mar 6 02:41:00.777645 systemd[1]: session-18.scope: Deactivated successfully. Mar 6 02:41:00.786207 systemd-logind[1544]: Session 18 logged out. Waiting for processes to exit. Mar 6 02:41:00.793159 systemd-logind[1544]: Removed session 18. Mar 6 02:41:05.798755 systemd[1]: Started sshd@18-10.0.0.69:22-10.0.0.1:36028.service - OpenSSH per-connection server daemon (10.0.0.1:36028). Mar 6 02:41:06.026794 sshd[6073]: Accepted publickey for core from 10.0.0.1 port 36028 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:06.032869 sshd-session[6073]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:06.042071 systemd-logind[1544]: New session 19 of user core. Mar 6 02:41:06.055809 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 6 02:41:06.414478 sshd[6076]: Connection closed by 10.0.0.1 port 36028 Mar 6 02:41:06.411247 sshd-session[6073]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:06.430876 systemd[1]: sshd@18-10.0.0.69:22-10.0.0.1:36028.service: Deactivated successfully. Mar 6 02:41:06.437171 systemd[1]: session-19.scope: Deactivated successfully. Mar 6 02:41:06.442149 systemd-logind[1544]: Session 19 logged out. Waiting for processes to exit. Mar 6 02:41:06.446843 systemd[1]: Started sshd@19-10.0.0.69:22-10.0.0.1:36040.service - OpenSSH per-connection server daemon (10.0.0.1:36040). Mar 6 02:41:06.456014 systemd-logind[1544]: Removed session 19. Mar 6 02:41:06.592807 sshd[6090]: Accepted publickey for core from 10.0.0.1 port 36040 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:06.595252 sshd-session[6090]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:06.651798 systemd-logind[1544]: New session 20 of user core. Mar 6 02:41:06.667997 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 6 02:41:07.460627 sshd[6093]: Connection closed by 10.0.0.1 port 36040 Mar 6 02:41:07.464668 sshd-session[6090]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:07.500066 systemd[1]: sshd@19-10.0.0.69:22-10.0.0.1:36040.service: Deactivated successfully. Mar 6 02:41:07.520953 systemd[1]: session-20.scope: Deactivated successfully. Mar 6 02:41:07.534493 systemd-logind[1544]: Session 20 logged out. Waiting for processes to exit. Mar 6 02:41:07.558767 systemd[1]: Started sshd@20-10.0.0.69:22-10.0.0.1:36054.service - OpenSSH per-connection server daemon (10.0.0.1:36054). Mar 6 02:41:07.573249 systemd-logind[1544]: Removed session 20. Mar 6 02:41:07.946863 sshd[6105]: Accepted publickey for core from 10.0.0.1 port 36054 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:07.951743 sshd-session[6105]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:07.961190 systemd-logind[1544]: New session 21 of user core. Mar 6 02:41:07.993776 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 6 02:41:08.978806 sshd[6108]: Connection closed by 10.0.0.1 port 36054 Mar 6 02:41:08.983062 sshd-session[6105]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:09.007887 systemd[1]: sshd@20-10.0.0.69:22-10.0.0.1:36054.service: Deactivated successfully. Mar 6 02:41:09.041961 systemd[1]: session-21.scope: Deactivated successfully. Mar 6 02:41:09.046857 systemd-logind[1544]: Session 21 logged out. Waiting for processes to exit. Mar 6 02:41:09.054836 systemd-logind[1544]: Removed session 21. Mar 6 02:41:14.006856 systemd[1]: Started sshd@21-10.0.0.69:22-10.0.0.1:59102.service - OpenSSH per-connection server daemon (10.0.0.1:59102). Mar 6 02:41:14.234720 sshd[6160]: Accepted publickey for core from 10.0.0.1 port 59102 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:14.238144 sshd-session[6160]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:14.250748 systemd-logind[1544]: New session 22 of user core. Mar 6 02:41:14.267188 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 6 02:41:14.909452 sshd[6169]: Connection closed by 10.0.0.1 port 59102 Mar 6 02:41:14.909691 sshd-session[6160]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:14.940113 systemd[1]: sshd@21-10.0.0.69:22-10.0.0.1:59102.service: Deactivated successfully. Mar 6 02:41:14.946729 systemd[1]: session-22.scope: Deactivated successfully. Mar 6 02:41:14.952192 systemd-logind[1544]: Session 22 logged out. Waiting for processes to exit. Mar 6 02:41:14.956272 systemd-logind[1544]: Removed session 22. Mar 6 02:41:16.271531 kubelet[2879]: E0306 02:41:16.271059 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:19.936489 systemd[1]: Started sshd@22-10.0.0.69:22-10.0.0.1:59104.service - OpenSSH per-connection server daemon (10.0.0.1:59104). Mar 6 02:41:20.161992 sshd[6212]: Accepted publickey for core from 10.0.0.1 port 59104 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:20.167053 sshd-session[6212]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:20.187101 systemd-logind[1544]: New session 23 of user core. Mar 6 02:41:20.204921 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 6 02:41:20.680184 sshd[6215]: Connection closed by 10.0.0.1 port 59104 Mar 6 02:41:20.681209 sshd-session[6212]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:20.695520 systemd[1]: sshd@22-10.0.0.69:22-10.0.0.1:59104.service: Deactivated successfully. Mar 6 02:41:20.701801 systemd[1]: session-23.scope: Deactivated successfully. Mar 6 02:41:20.705805 systemd-logind[1544]: Session 23 logged out. Waiting for processes to exit. Mar 6 02:41:20.711780 systemd-logind[1544]: Removed session 23. Mar 6 02:41:22.282431 kubelet[2879]: E0306 02:41:22.282199 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:23.272000 kubelet[2879]: E0306 02:41:23.271673 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:23.276927 kubelet[2879]: E0306 02:41:23.276844 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:25.722176 systemd[1]: Started sshd@23-10.0.0.69:22-10.0.0.1:44020.service - OpenSSH per-connection server daemon (10.0.0.1:44020). Mar 6 02:41:25.938166 sshd[6274]: Accepted publickey for core from 10.0.0.1 port 44020 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:25.944488 sshd-session[6274]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:25.974072 systemd-logind[1544]: New session 24 of user core. Mar 6 02:41:25.981940 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 6 02:41:26.322272 sshd[6277]: Connection closed by 10.0.0.1 port 44020 Mar 6 02:41:26.324204 sshd-session[6274]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:26.345161 systemd[1]: sshd@23-10.0.0.69:22-10.0.0.1:44020.service: Deactivated successfully. Mar 6 02:41:26.361842 systemd[1]: session-24.scope: Deactivated successfully. Mar 6 02:41:26.366493 systemd-logind[1544]: Session 24 logged out. Waiting for processes to exit. Mar 6 02:41:26.375899 systemd-logind[1544]: Removed session 24. Mar 6 02:41:28.279096 kubelet[2879]: E0306 02:41:28.278992 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:31.343104 systemd[1]: Started sshd@24-10.0.0.69:22-10.0.0.1:44024.service - OpenSSH per-connection server daemon (10.0.0.1:44024). Mar 6 02:41:31.466235 sshd[6310]: Accepted publickey for core from 10.0.0.1 port 44024 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:31.469541 sshd-session[6310]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:31.484563 systemd-logind[1544]: New session 25 of user core. Mar 6 02:41:31.492798 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 6 02:41:31.808189 sshd[6313]: Connection closed by 10.0.0.1 port 44024 Mar 6 02:41:31.808762 sshd-session[6310]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:31.825922 systemd[1]: sshd@24-10.0.0.69:22-10.0.0.1:44024.service: Deactivated successfully. Mar 6 02:41:31.831866 systemd[1]: session-25.scope: Deactivated successfully. Mar 6 02:41:31.837752 systemd-logind[1544]: Session 25 logged out. Waiting for processes to exit. Mar 6 02:41:31.849999 systemd[1]: Started sshd@25-10.0.0.69:22-10.0.0.1:44038.service - OpenSSH per-connection server daemon (10.0.0.1:44038). Mar 6 02:41:31.855451 systemd-logind[1544]: Removed session 25. Mar 6 02:41:31.975147 sshd[6331]: Accepted publickey for core from 10.0.0.1 port 44038 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:31.978938 sshd-session[6331]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:31.997733 systemd-logind[1544]: New session 26 of user core. Mar 6 02:41:32.005216 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 6 02:41:33.051850 sshd[6334]: Connection closed by 10.0.0.1 port 44038 Mar 6 02:41:33.055944 sshd-session[6331]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:33.079569 systemd[1]: sshd@25-10.0.0.69:22-10.0.0.1:44038.service: Deactivated successfully. Mar 6 02:41:33.086005 systemd[1]: session-26.scope: Deactivated successfully. Mar 6 02:41:33.092027 systemd-logind[1544]: Session 26 logged out. Waiting for processes to exit. Mar 6 02:41:33.101832 systemd[1]: Started sshd@26-10.0.0.69:22-10.0.0.1:35536.service - OpenSSH per-connection server daemon (10.0.0.1:35536). Mar 6 02:41:33.112757 systemd-logind[1544]: Removed session 26. Mar 6 02:41:48.895902 sshd[6345]: Accepted publickey for core from 10.0.0.1 port 35536 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:49.853424 sshd-session[6345]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:50.045925 systemd-logind[1544]: New session 27 of user core. Mar 6 02:41:50.057653 systemd[1]: Started session-27.scope - Session 27 of User core. Mar 6 02:41:50.140619 systemd[1]: cri-containerd-abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075.scope: Deactivated successfully. Mar 6 02:41:50.141997 systemd[1]: cri-containerd-abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075.scope: Consumed 8.856s CPU time, 34.2M memory peak, 13.6M read from disk. Mar 6 02:41:50.229492 containerd[1569]: time="2026-03-06T02:41:50.228744604Z" level=info msg="received container exit event container_id:\"abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075\" id:\"abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075\" pid:2735 exit_status:1 exited_at:{seconds:1772764910 nanos:190994797}" Mar 6 02:41:50.727429 kubelet[2879]: E0306 02:41:50.715963 2879 kubelet.go:2691] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="16.037s" Mar 6 02:41:50.789640 kubelet[2879]: E0306 02:41:50.789570 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:50.791967 kubelet[2879]: E0306 02:41:50.791801 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:51.259086 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075-rootfs.mount: Deactivated successfully. Mar 6 02:41:51.833056 kubelet[2879]: I0306 02:41:51.832238 2879 scope.go:122] "RemoveContainer" containerID="abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075" Mar 6 02:41:51.835445 kubelet[2879]: E0306 02:41:51.834647 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:51.947365 containerd[1569]: time="2026-03-06T02:41:51.946767438Z" level=info msg="CreateContainer within sandbox \"76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Mar 6 02:41:52.054161 containerd[1569]: time="2026-03-06T02:41:52.053544889Z" level=info msg="Container 36bd3091de6af0e8570cc3fac397b99e9d7bd8c554dfbac751a93750b20909d4: CDI devices from CRI Config.CDIDevices: []" Mar 6 02:41:52.066137 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2581851931.mount: Deactivated successfully. Mar 6 02:41:52.169246 containerd[1569]: time="2026-03-06T02:41:52.167530529Z" level=info msg="CreateContainer within sandbox \"76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"36bd3091de6af0e8570cc3fac397b99e9d7bd8c554dfbac751a93750b20909d4\"" Mar 6 02:41:52.171626 containerd[1569]: time="2026-03-06T02:41:52.171215311Z" level=info msg="StartContainer for \"36bd3091de6af0e8570cc3fac397b99e9d7bd8c554dfbac751a93750b20909d4\"" Mar 6 02:41:52.187612 containerd[1569]: time="2026-03-06T02:41:52.187558812Z" level=info msg="connecting to shim 36bd3091de6af0e8570cc3fac397b99e9d7bd8c554dfbac751a93750b20909d4" address="unix:///run/containerd/s/34a31aeda2ae590da7b89310a1b259a615a310a1e045c82ab392d45f799ab99c" protocol=ttrpc version=3 Mar 6 02:41:52.416817 systemd[1]: Started cri-containerd-36bd3091de6af0e8570cc3fac397b99e9d7bd8c554dfbac751a93750b20909d4.scope - libcontainer container 36bd3091de6af0e8570cc3fac397b99e9d7bd8c554dfbac751a93750b20909d4. Mar 6 02:41:52.886974 containerd[1569]: time="2026-03-06T02:41:52.886574605Z" level=info msg="StartContainer for \"36bd3091de6af0e8570cc3fac397b99e9d7bd8c554dfbac751a93750b20909d4\" returns successfully" Mar 6 02:41:53.081844 sshd[6349]: Connection closed by 10.0.0.1 port 35536 Mar 6 02:41:53.082611 sshd-session[6345]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:53.098152 systemd[1]: sshd@26-10.0.0.69:22-10.0.0.1:35536.service: Deactivated successfully. Mar 6 02:41:53.098894 systemd[1]: sshd@26-10.0.0.69:22-10.0.0.1:35536.service: Consumed 1.280s CPU time, 5.2M memory peak. Mar 6 02:41:53.104025 systemd[1]: session-27.scope: Deactivated successfully. Mar 6 02:41:53.105649 systemd[1]: session-27.scope: Consumed 1.267s CPU time, 43.6M memory peak. Mar 6 02:41:53.110821 systemd-logind[1544]: Session 27 logged out. Waiting for processes to exit. Mar 6 02:41:53.114888 systemd[1]: Started sshd@27-10.0.0.69:22-10.0.0.1:47758.service - OpenSSH per-connection server daemon (10.0.0.1:47758). Mar 6 02:41:53.146627 systemd-logind[1544]: Removed session 27. Mar 6 02:41:53.296276 sshd[6512]: Accepted publickey for core from 10.0.0.1 port 47758 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:53.310066 sshd-session[6512]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:53.339645 systemd-logind[1544]: New session 28 of user core. Mar 6 02:41:53.354232 systemd[1]: Started session-28.scope - Session 28 of User core. Mar 6 02:41:53.961518 kubelet[2879]: E0306 02:41:53.961199 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:54.756590 sshd[6520]: Connection closed by 10.0.0.1 port 47758 Mar 6 02:41:54.757609 sshd-session[6512]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:54.780257 systemd[1]: sshd@27-10.0.0.69:22-10.0.0.1:47758.service: Deactivated successfully. Mar 6 02:41:54.792982 systemd[1]: session-28.scope: Deactivated successfully. Mar 6 02:41:54.799767 systemd-logind[1544]: Session 28 logged out. Waiting for processes to exit. Mar 6 02:41:54.809029 systemd[1]: Started sshd@28-10.0.0.69:22-10.0.0.1:47774.service - OpenSSH per-connection server daemon (10.0.0.1:47774). Mar 6 02:41:54.813515 systemd-logind[1544]: Removed session 28. Mar 6 02:41:54.965257 kubelet[2879]: E0306 02:41:54.964953 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:41:55.053190 sshd[6567]: Accepted publickey for core from 10.0.0.1 port 47774 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:41:55.056487 sshd-session[6567]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:41:55.074718 systemd-logind[1544]: New session 29 of user core. Mar 6 02:41:55.080754 systemd[1]: Started session-29.scope - Session 29 of User core. Mar 6 02:41:55.389759 sshd[6570]: Connection closed by 10.0.0.1 port 47774 Mar 6 02:41:55.391522 sshd-session[6567]: pam_unix(sshd:session): session closed for user core Mar 6 02:41:55.400191 systemd[1]: sshd@28-10.0.0.69:22-10.0.0.1:47774.service: Deactivated successfully. Mar 6 02:41:55.405738 systemd[1]: session-29.scope: Deactivated successfully. Mar 6 02:41:55.407904 systemd-logind[1544]: Session 29 logged out. Waiting for processes to exit. Mar 6 02:41:55.411582 systemd-logind[1544]: Removed session 29. Mar 6 02:42:00.403101 kubelet[2879]: E0306 02:42:00.402577 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:42:00.425235 systemd[1]: Started sshd@29-10.0.0.69:22-10.0.0.1:47782.service - OpenSSH per-connection server daemon (10.0.0.1:47782). Mar 6 02:42:00.578753 sshd[6586]: Accepted publickey for core from 10.0.0.1 port 47782 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:42:00.592821 sshd-session[6586]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:42:00.631109 systemd-logind[1544]: New session 30 of user core. Mar 6 02:42:00.651179 systemd[1]: Started session-30.scope - Session 30 of User core. Mar 6 02:42:00.968921 sshd[6589]: Connection closed by 10.0.0.1 port 47782 Mar 6 02:42:00.972516 sshd-session[6586]: pam_unix(sshd:session): session closed for user core Mar 6 02:42:00.986787 systemd[1]: sshd@29-10.0.0.69:22-10.0.0.1:47782.service: Deactivated successfully. Mar 6 02:42:00.993615 systemd[1]: session-30.scope: Deactivated successfully. Mar 6 02:42:01.003032 systemd-logind[1544]: Session 30 logged out. Waiting for processes to exit. Mar 6 02:42:01.012118 systemd-logind[1544]: Removed session 30. Mar 6 02:42:05.992492 systemd[1]: Started sshd@30-10.0.0.69:22-10.0.0.1:60470.service - OpenSSH per-connection server daemon (10.0.0.1:60470). Mar 6 02:42:06.102042 sshd[6606]: Accepted publickey for core from 10.0.0.1 port 60470 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:42:06.105265 sshd-session[6606]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:42:06.127029 systemd-logind[1544]: New session 31 of user core. Mar 6 02:42:06.135040 systemd[1]: Started session-31.scope - Session 31 of User core. Mar 6 02:42:06.365239 sshd[6609]: Connection closed by 10.0.0.1 port 60470 Mar 6 02:42:06.366534 sshd-session[6606]: pam_unix(sshd:session): session closed for user core Mar 6 02:42:06.373485 systemd[1]: sshd@30-10.0.0.69:22-10.0.0.1:60470.service: Deactivated successfully. Mar 6 02:42:06.378870 systemd[1]: session-31.scope: Deactivated successfully. Mar 6 02:42:06.391985 systemd-logind[1544]: Session 31 logged out. Waiting for processes to exit. Mar 6 02:42:06.395677 systemd-logind[1544]: Removed session 31. Mar 6 02:42:10.423266 kubelet[2879]: E0306 02:42:10.422907 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:42:11.070884 kubelet[2879]: E0306 02:42:11.070684 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:42:11.193391 containerd[1569]: time="2026-03-06T02:42:11.131228158Z" level=warning msg="container event discarded" container=a85e44ca8b894f7dcb7b4bf2b032a3e44a8bf6de4907c99ea6a9974e4e12bba9 type=CONTAINER_CREATED_EVENT Mar 6 02:42:11.302517 containerd[1569]: time="2026-03-06T02:42:11.302144916Z" level=warning msg="container event discarded" container=a85e44ca8b894f7dcb7b4bf2b032a3e44a8bf6de4907c99ea6a9974e4e12bba9 type=CONTAINER_STARTED_EVENT Mar 6 02:42:11.302517 containerd[1569]: time="2026-03-06T02:42:11.302217270Z" level=warning msg="container event discarded" container=741c5061cf0493872b8c57c8cfa830ad32da9b3f36c45f395b83146dce45632e type=CONTAINER_CREATED_EVENT Mar 6 02:42:11.302517 containerd[1569]: time="2026-03-06T02:42:11.302232298Z" level=warning msg="container event discarded" container=741c5061cf0493872b8c57c8cfa830ad32da9b3f36c45f395b83146dce45632e type=CONTAINER_STARTED_EVENT Mar 6 02:42:11.302517 containerd[1569]: time="2026-03-06T02:42:11.302243770Z" level=warning msg="container event discarded" container=76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299 type=CONTAINER_CREATED_EVENT Mar 6 02:42:11.302517 containerd[1569]: time="2026-03-06T02:42:11.302256644Z" level=warning msg="container event discarded" container=76b99eb9aa17d222f7ba88b8c6c10eee1c7bb4be737530f604150a00a0ff2299 type=CONTAINER_STARTED_EVENT Mar 6 02:42:11.379070 containerd[1569]: time="2026-03-06T02:42:11.378213938Z" level=warning msg="container event discarded" container=bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd type=CONTAINER_CREATED_EVENT Mar 6 02:42:11.414822 systemd[1]: Started sshd@31-10.0.0.69:22-10.0.0.1:60476.service - OpenSSH per-connection server daemon (10.0.0.1:60476). Mar 6 02:42:11.432434 containerd[1569]: time="2026-03-06T02:42:11.431060485Z" level=warning msg="container event discarded" container=2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695 type=CONTAINER_CREATED_EVENT Mar 6 02:42:11.495569 containerd[1569]: time="2026-03-06T02:42:11.494904642Z" level=warning msg="container event discarded" container=abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075 type=CONTAINER_CREATED_EVENT Mar 6 02:42:11.545892 sshd[6623]: Accepted publickey for core from 10.0.0.1 port 60476 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:42:11.550259 sshd-session[6623]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:42:11.567512 systemd-logind[1544]: New session 32 of user core. Mar 6 02:42:11.576080 systemd[1]: Started session-32.scope - Session 32 of User core. Mar 6 02:42:11.824018 sshd[6626]: Connection closed by 10.0.0.1 port 60476 Mar 6 02:42:11.825034 sshd-session[6623]: pam_unix(sshd:session): session closed for user core Mar 6 02:42:11.834190 systemd[1]: sshd@31-10.0.0.69:22-10.0.0.1:60476.service: Deactivated successfully. Mar 6 02:42:11.840848 systemd[1]: session-32.scope: Deactivated successfully. Mar 6 02:42:11.843913 systemd-logind[1544]: Session 32 logged out. Waiting for processes to exit. Mar 6 02:42:11.851029 systemd-logind[1544]: Removed session 32. Mar 6 02:42:11.863536 containerd[1569]: time="2026-03-06T02:42:11.863241885Z" level=warning msg="container event discarded" container=bb7ba851b2d15b27ab209e59e28bf3d488d4f4ef13a5965647ecc757160a46fd type=CONTAINER_STARTED_EVENT Mar 6 02:42:12.028444 containerd[1569]: time="2026-03-06T02:42:12.027267697Z" level=warning msg="container event discarded" container=2d0b3a21036aff1075dc60d72dc7e2a13b060574be2ecdcf6874e745fe096695 type=CONTAINER_STARTED_EVENT Mar 6 02:42:12.028444 containerd[1569]: time="2026-03-06T02:42:12.027423317Z" level=warning msg="container event discarded" container=abe9046a28317ac1e0ba577656184c453e7185ba8a47bcd9929c16db334f1075 type=CONTAINER_STARTED_EVENT Mar 6 02:42:16.848262 systemd[1]: Started sshd@32-10.0.0.69:22-10.0.0.1:40356.service - OpenSSH per-connection server daemon (10.0.0.1:40356). Mar 6 02:42:16.988506 sshd[6687]: Accepted publickey for core from 10.0.0.1 port 40356 ssh2: RSA SHA256:ScMF4t+sRFLe42Axw5QjqGy4QurXMGM75Y6m1mn+/uU Mar 6 02:42:16.995156 sshd-session[6687]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 02:42:17.032575 systemd-logind[1544]: New session 33 of user core. Mar 6 02:42:17.040505 systemd[1]: Started session-33.scope - Session 33 of User core. Mar 6 02:42:17.270205 kubelet[2879]: E0306 02:42:17.269660 2879 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 02:42:17.340216 sshd[6690]: Connection closed by 10.0.0.1 port 40356 Mar 6 02:42:17.340658 sshd-session[6687]: pam_unix(sshd:session): session closed for user core Mar 6 02:42:17.348596 systemd[1]: sshd@32-10.0.0.69:22-10.0.0.1:40356.service: Deactivated successfully. Mar 6 02:42:17.354554 systemd[1]: session-33.scope: Deactivated successfully. Mar 6 02:42:17.357697 systemd-logind[1544]: Session 33 logged out. Waiting for processes to exit. Mar 6 02:42:17.361969 systemd-logind[1544]: Removed session 33.