Mar 2 14:25:40.764051 kernel: Linux version 6.12.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Mon Mar 2 10:28:24 -00 2026 Mar 2 14:25:40.764122 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=82731586f036a8515942386c762f58de23efa7b4e7ecf4198e267e112154cbc2 Mar 2 14:25:40.764136 kernel: BIOS-provided physical RAM map: Mar 2 14:25:40.764149 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 2 14:25:40.764158 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Mar 2 14:25:40.764167 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Mar 2 14:25:40.764178 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Mar 2 14:25:40.764187 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Mar 2 14:25:40.764197 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Mar 2 14:25:40.764206 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Mar 2 14:25:40.764216 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Mar 2 14:25:40.764225 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Mar 2 14:25:40.764999 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Mar 2 14:25:40.765013 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Mar 2 14:25:40.765049 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Mar 2 14:25:40.765060 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Mar 2 14:25:40.765070 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Mar 2 14:25:40.765086 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Mar 2 14:25:40.765097 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Mar 2 14:25:40.765105 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Mar 2 14:25:40.765114 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Mar 2 14:25:40.765124 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Mar 2 14:25:40.765136 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Mar 2 14:25:40.765144 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 2 14:25:40.765153 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Mar 2 14:25:40.765162 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 2 14:25:40.765174 kernel: NX (Execute Disable) protection: active Mar 2 14:25:40.765182 kernel: APIC: Static calls initialized Mar 2 14:25:40.765195 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Mar 2 14:25:40.765207 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Mar 2 14:25:40.765216 kernel: extended physical RAM map: Mar 2 14:25:40.765225 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 2 14:25:40.765306 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Mar 2 14:25:40.765318 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Mar 2 14:25:40.765329 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Mar 2 14:25:40.765340 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Mar 2 14:25:40.765350 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Mar 2 14:25:40.765359 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Mar 2 14:25:40.765367 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Mar 2 14:25:40.765380 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Mar 2 14:25:40.765393 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Mar 2 14:25:40.765402 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Mar 2 14:25:40.765411 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Mar 2 14:25:40.765420 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Mar 2 14:25:40.765431 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Mar 2 14:25:40.765440 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Mar 2 14:25:40.765449 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Mar 2 14:25:40.765458 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Mar 2 14:25:40.765466 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Mar 2 14:25:40.765475 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Mar 2 14:25:40.765485 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Mar 2 14:25:40.765495 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Mar 2 14:25:40.765506 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Mar 2 14:25:40.765516 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Mar 2 14:25:40.765527 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Mar 2 14:25:40.765540 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 2 14:25:40.765551 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Mar 2 14:25:40.765561 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 2 14:25:40.765571 kernel: efi: EFI v2.7 by EDK II Mar 2 14:25:40.765582 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Mar 2 14:25:40.765593 kernel: random: crng init done Mar 2 14:25:40.765603 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Mar 2 14:25:40.765613 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Mar 2 14:25:40.765624 kernel: secureboot: Secure boot disabled Mar 2 14:25:40.765634 kernel: SMBIOS 2.8 present. Mar 2 14:25:40.766385 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Mar 2 14:25:40.766404 kernel: DMI: Memory slots populated: 1/1 Mar 2 14:25:40.766415 kernel: Hypervisor detected: KVM Mar 2 14:25:40.766427 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Mar 2 14:25:40.766438 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 2 14:25:40.766448 kernel: kvm-clock: using sched offset of 11355307709 cycles Mar 2 14:25:40.766462 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 2 14:25:40.766473 kernel: tsc: Detected 2445.426 MHz processor Mar 2 14:25:40.766484 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 2 14:25:40.766495 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 2 14:25:40.766505 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Mar 2 14:25:40.766516 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Mar 2 14:25:40.766531 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 2 14:25:40.766542 kernel: Using GB pages for direct mapping Mar 2 14:25:40.766552 kernel: ACPI: Early table checksum verification disabled Mar 2 14:25:40.766563 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Mar 2 14:25:40.766574 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Mar 2 14:25:40.766585 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 14:25:40.766595 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 14:25:40.766606 kernel: ACPI: FACS 0x000000009CBDD000 000040 Mar 2 14:25:40.766617 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 14:25:40.766631 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 14:25:40.766642 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 14:25:40.766653 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 14:25:40.766663 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Mar 2 14:25:40.766674 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Mar 2 14:25:40.766685 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Mar 2 14:25:40.766695 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Mar 2 14:25:40.766706 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Mar 2 14:25:40.766745 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Mar 2 14:25:40.766755 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Mar 2 14:25:40.766768 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Mar 2 14:25:40.766779 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Mar 2 14:25:40.766789 kernel: No NUMA configuration found Mar 2 14:25:40.766800 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Mar 2 14:25:40.766811 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Mar 2 14:25:40.766821 kernel: Zone ranges: Mar 2 14:25:40.766832 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 2 14:25:40.766843 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Mar 2 14:25:40.766857 kernel: Normal empty Mar 2 14:25:40.766867 kernel: Device empty Mar 2 14:25:40.766878 kernel: Movable zone start for each node Mar 2 14:25:40.766888 kernel: Early memory node ranges Mar 2 14:25:40.766899 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Mar 2 14:25:40.766910 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Mar 2 14:25:40.766920 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Mar 2 14:25:40.766931 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Mar 2 14:25:40.766941 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Mar 2 14:25:40.766955 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Mar 2 14:25:40.766966 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Mar 2 14:25:40.766976 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Mar 2 14:25:40.766987 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Mar 2 14:25:40.766998 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 2 14:25:40.767022 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Mar 2 14:25:40.767037 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Mar 2 14:25:40.767048 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 2 14:25:40.767059 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Mar 2 14:25:40.767070 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Mar 2 14:25:40.767081 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Mar 2 14:25:40.767092 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Mar 2 14:25:40.767107 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Mar 2 14:25:40.767118 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 2 14:25:40.767129 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 2 14:25:40.767140 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 2 14:25:40.767152 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 2 14:25:40.767166 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 2 14:25:40.767177 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 2 14:25:40.767188 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 2 14:25:40.767199 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 2 14:25:40.767211 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 2 14:25:40.767222 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 2 14:25:40.767514 kernel: TSC deadline timer available Mar 2 14:25:40.767528 kernel: CPU topo: Max. logical packages: 1 Mar 2 14:25:40.767540 kernel: CPU topo: Max. logical dies: 1 Mar 2 14:25:40.767556 kernel: CPU topo: Max. dies per package: 1 Mar 2 14:25:40.767566 kernel: CPU topo: Max. threads per core: 1 Mar 2 14:25:40.767578 kernel: CPU topo: Num. cores per package: 4 Mar 2 14:25:40.767589 kernel: CPU topo: Num. threads per package: 4 Mar 2 14:25:40.767600 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Mar 2 14:25:40.767611 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 2 14:25:40.767623 kernel: kvm-guest: KVM setup pv remote TLB flush Mar 2 14:25:40.767634 kernel: kvm-guest: setup PV sched yield Mar 2 14:25:40.767646 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Mar 2 14:25:40.767660 kernel: Booting paravirtualized kernel on KVM Mar 2 14:25:40.767670 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 2 14:25:40.767681 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Mar 2 14:25:40.767691 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Mar 2 14:25:40.767700 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Mar 2 14:25:40.767709 kernel: pcpu-alloc: [0] 0 1 2 3 Mar 2 14:25:40.767718 kernel: kvm-guest: PV spinlocks enabled Mar 2 14:25:40.767728 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Mar 2 14:25:40.767741 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=82731586f036a8515942386c762f58de23efa7b4e7ecf4198e267e112154cbc2 Mar 2 14:25:40.767757 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 2 14:25:40.767768 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 2 14:25:40.767781 kernel: Fallback order for Node 0: 0 Mar 2 14:25:40.767791 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Mar 2 14:25:40.767802 kernel: Policy zone: DMA32 Mar 2 14:25:40.767811 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 2 14:25:40.767820 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Mar 2 14:25:40.767830 kernel: ftrace: allocating 40099 entries in 157 pages Mar 2 14:25:40.767844 kernel: ftrace: allocated 157 pages with 5 groups Mar 2 14:25:40.767856 kernel: Dynamic Preempt: voluntary Mar 2 14:25:40.767867 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 2 14:25:40.767885 kernel: rcu: RCU event tracing is enabled. Mar 2 14:25:40.767897 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Mar 2 14:25:40.767908 kernel: Trampoline variant of Tasks RCU enabled. Mar 2 14:25:40.767919 kernel: Rude variant of Tasks RCU enabled. Mar 2 14:25:40.767931 kernel: Tracing variant of Tasks RCU enabled. Mar 2 14:25:40.767942 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 2 14:25:40.767953 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Mar 2 14:25:40.767967 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 14:25:40.767979 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 14:25:40.767990 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 14:25:40.768001 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Mar 2 14:25:40.768013 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 2 14:25:40.768024 kernel: Console: colour dummy device 80x25 Mar 2 14:25:40.768035 kernel: printk: legacy console [ttyS0] enabled Mar 2 14:25:40.768046 kernel: ACPI: Core revision 20240827 Mar 2 14:25:40.768058 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 2 14:25:40.768072 kernel: APIC: Switch to symmetric I/O mode setup Mar 2 14:25:40.768083 kernel: x2apic enabled Mar 2 14:25:40.768094 kernel: APIC: Switched APIC routing to: physical x2apic Mar 2 14:25:40.768105 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Mar 2 14:25:40.768117 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Mar 2 14:25:40.768128 kernel: kvm-guest: setup PV IPIs Mar 2 14:25:40.768140 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 2 14:25:40.768151 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Mar 2 14:25:40.768165 kernel: Calibrating delay loop (skipped) preset value.. 4890.85 BogoMIPS (lpj=2445426) Mar 2 14:25:40.768176 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 2 14:25:40.768188 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Mar 2 14:25:40.768199 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Mar 2 14:25:40.768210 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 2 14:25:40.768222 kernel: Spectre V2 : Mitigation: Retpolines Mar 2 14:25:40.768302 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Mar 2 14:25:40.768313 kernel: Speculative Store Bypass: Vulnerable Mar 2 14:25:40.768322 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Mar 2 14:25:40.768341 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Mar 2 14:25:40.768351 kernel: active return thunk: srso_alias_return_thunk Mar 2 14:25:40.768361 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Mar 2 14:25:40.768372 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Mar 2 14:25:40.768384 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Mar 2 14:25:40.768395 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 2 14:25:40.768404 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 2 14:25:40.768417 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 2 14:25:40.768428 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 2 14:25:40.768444 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Mar 2 14:25:40.768453 kernel: Freeing SMP alternatives memory: 32K Mar 2 14:25:40.768465 kernel: pid_max: default: 32768 minimum: 301 Mar 2 14:25:40.768478 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Mar 2 14:25:40.768487 kernel: landlock: Up and running. Mar 2 14:25:40.768498 kernel: SELinux: Initializing. Mar 2 14:25:40.768510 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 14:25:40.768520 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 14:25:40.768531 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Mar 2 14:25:40.768547 kernel: Performance Events: PMU not available due to virtualization, using software events only. Mar 2 14:25:40.768557 kernel: signal: max sigframe size: 1776 Mar 2 14:25:40.768568 kernel: rcu: Hierarchical SRCU implementation. Mar 2 14:25:40.768579 kernel: rcu: Max phase no-delay instances is 400. Mar 2 14:25:40.768590 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Mar 2 14:25:40.768600 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Mar 2 14:25:40.768610 kernel: smp: Bringing up secondary CPUs ... Mar 2 14:25:40.768621 kernel: smpboot: x86: Booting SMP configuration: Mar 2 14:25:40.768631 kernel: .... node #0, CPUs: #1 #2 #3 Mar 2 14:25:40.768645 kernel: smp: Brought up 1 node, 4 CPUs Mar 2 14:25:40.768656 kernel: smpboot: Total of 4 processors activated (19563.40 BogoMIPS) Mar 2 14:25:40.768667 kernel: Memory: 2414476K/2565800K available (14336K kernel code, 2445K rwdata, 26064K rodata, 46192K init, 2568K bss, 145388K reserved, 0K cma-reserved) Mar 2 14:25:40.768677 kernel: devtmpfs: initialized Mar 2 14:25:40.768688 kernel: x86/mm: Memory block size: 128MB Mar 2 14:25:40.768698 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Mar 2 14:25:40.768709 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Mar 2 14:25:40.768720 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Mar 2 14:25:40.768735 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Mar 2 14:25:40.768747 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Mar 2 14:25:40.768757 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Mar 2 14:25:40.768767 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 2 14:25:40.768778 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Mar 2 14:25:40.768789 kernel: pinctrl core: initialized pinctrl subsystem Mar 2 14:25:40.768799 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 2 14:25:40.768810 kernel: audit: initializing netlink subsys (disabled) Mar 2 14:25:40.768820 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 2 14:25:40.768834 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 2 14:25:40.768845 kernel: audit: type=2000 audit(1772461531.368:1): state=initialized audit_enabled=0 res=1 Mar 2 14:25:40.768855 kernel: cpuidle: using governor menu Mar 2 14:25:40.768866 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 2 14:25:40.768876 kernel: dca service started, version 1.12.1 Mar 2 14:25:40.768887 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Mar 2 14:25:40.768897 kernel: PCI: Using configuration type 1 for base access Mar 2 14:25:40.768908 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 2 14:25:40.768919 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 2 14:25:40.768932 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 2 14:25:40.768943 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 2 14:25:40.768953 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 2 14:25:40.768965 kernel: ACPI: Added _OSI(Module Device) Mar 2 14:25:40.768976 kernel: ACPI: Added _OSI(Processor Device) Mar 2 14:25:40.768986 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 2 14:25:40.768997 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 2 14:25:40.769010 kernel: ACPI: Interpreter enabled Mar 2 14:25:40.769019 kernel: ACPI: PM: (supports S0 S3 S5) Mar 2 14:25:40.769033 kernel: ACPI: Using IOAPIC for interrupt routing Mar 2 14:25:40.769046 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 2 14:25:40.769056 kernel: PCI: Using E820 reservations for host bridge windows Mar 2 14:25:40.769065 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 2 14:25:40.769076 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 2 14:25:40.770461 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 2 14:25:40.770651 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Mar 2 14:25:40.770861 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Mar 2 14:25:40.770885 kernel: PCI host bridge to bus 0000:00 Mar 2 14:25:40.771061 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 2 14:25:40.771221 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 2 14:25:40.772136 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 2 14:25:40.772371 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Mar 2 14:25:40.772639 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Mar 2 14:25:40.772836 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Mar 2 14:25:40.773154 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 2 14:25:40.775650 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Mar 2 14:25:40.775894 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Mar 2 14:25:40.776075 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Mar 2 14:25:40.776322 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Mar 2 14:25:40.776516 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Mar 2 14:25:40.776710 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 2 14:25:40.776910 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Mar 2 14:25:40.777086 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Mar 2 14:25:40.777660 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Mar 2 14:25:40.777848 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Mar 2 14:25:40.778024 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Mar 2 14:25:40.778188 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Mar 2 14:25:40.778455 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Mar 2 14:25:40.778628 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Mar 2 14:25:40.778819 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Mar 2 14:25:40.778985 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Mar 2 14:25:40.779156 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Mar 2 14:25:40.779394 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Mar 2 14:25:40.779570 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Mar 2 14:25:40.779754 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Mar 2 14:25:40.779936 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 2 14:25:40.780109 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Mar 2 14:25:40.781602 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Mar 2 14:25:40.781836 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Mar 2 14:25:40.782036 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Mar 2 14:25:40.782304 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Mar 2 14:25:40.782323 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 2 14:25:40.782336 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 2 14:25:40.782347 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 2 14:25:40.782359 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 2 14:25:40.782370 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 2 14:25:40.783386 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 2 14:25:40.783404 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 2 14:25:40.783421 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 2 14:25:40.783432 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 2 14:25:40.783442 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 2 14:25:40.783453 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 2 14:25:40.783465 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 2 14:25:40.783476 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 2 14:25:40.783488 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 2 14:25:40.783499 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 2 14:25:40.783510 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 2 14:25:40.783525 kernel: iommu: Default domain type: Translated Mar 2 14:25:40.783536 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 2 14:25:40.783547 kernel: efivars: Registered efivars operations Mar 2 14:25:40.783560 kernel: PCI: Using ACPI for IRQ routing Mar 2 14:25:40.783573 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 2 14:25:40.783586 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Mar 2 14:25:40.783596 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Mar 2 14:25:40.783605 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Mar 2 14:25:40.783616 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Mar 2 14:25:40.783629 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Mar 2 14:25:40.783645 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Mar 2 14:25:40.783655 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Mar 2 14:25:40.783666 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Mar 2 14:25:40.783862 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 2 14:25:40.785495 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 2 14:25:40.785666 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 2 14:25:40.785681 kernel: vgaarb: loaded Mar 2 14:25:40.785699 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 2 14:25:40.785710 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 2 14:25:40.785722 kernel: clocksource: Switched to clocksource kvm-clock Mar 2 14:25:40.785734 kernel: VFS: Disk quotas dquot_6.6.0 Mar 2 14:25:40.785747 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 2 14:25:40.785758 kernel: pnp: PnP ACPI init Mar 2 14:25:40.786036 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Mar 2 14:25:40.786056 kernel: pnp: PnP ACPI: found 6 devices Mar 2 14:25:40.786068 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 2 14:25:40.786084 kernel: NET: Registered PF_INET protocol family Mar 2 14:25:40.786095 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 2 14:25:40.786107 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 2 14:25:40.786118 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 2 14:25:40.786130 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 2 14:25:40.786163 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 2 14:25:40.786178 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 2 14:25:40.786190 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 14:25:40.786205 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 14:25:40.786217 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 2 14:25:40.788501 kernel: NET: Registered PF_XDP protocol family Mar 2 14:25:40.788697 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Mar 2 14:25:40.788881 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Mar 2 14:25:40.789044 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 2 14:25:40.789191 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 2 14:25:40.789430 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 2 14:25:40.789591 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Mar 2 14:25:40.789743 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Mar 2 14:25:40.789901 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Mar 2 14:25:40.789918 kernel: PCI: CLS 0 bytes, default 64 Mar 2 14:25:40.789930 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Mar 2 14:25:40.789942 kernel: Initialise system trusted keyrings Mar 2 14:25:40.789954 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 2 14:25:40.789966 kernel: Key type asymmetric registered Mar 2 14:25:40.789978 kernel: Asymmetric key parser 'x509' registered Mar 2 14:25:40.789994 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Mar 2 14:25:40.790006 kernel: io scheduler mq-deadline registered Mar 2 14:25:40.790018 kernel: io scheduler kyber registered Mar 2 14:25:40.790030 kernel: io scheduler bfq registered Mar 2 14:25:40.790042 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 2 14:25:40.790055 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 2 14:25:40.790067 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 2 14:25:40.790079 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Mar 2 14:25:40.790094 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 2 14:25:40.790106 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 2 14:25:40.790118 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 2 14:25:40.790133 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 2 14:25:40.790145 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 2 14:25:40.790840 kernel: rtc_cmos 00:04: RTC can wake from S4 Mar 2 14:25:40.790867 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Mar 2 14:25:40.791024 kernel: rtc_cmos 00:04: registered as rtc0 Mar 2 14:25:40.791177 kernel: rtc_cmos 00:04: setting system clock to 2026-03-02T14:25:39 UTC (1772461539) Mar 2 14:25:40.791471 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Mar 2 14:25:40.791490 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Mar 2 14:25:40.791502 kernel: efifb: probing for efifb Mar 2 14:25:40.791514 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Mar 2 14:25:40.791525 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Mar 2 14:25:40.791541 kernel: efifb: scrolling: redraw Mar 2 14:25:40.791553 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 2 14:25:40.791565 kernel: Console: switching to colour frame buffer device 160x50 Mar 2 14:25:40.791576 kernel: fb0: EFI VGA frame buffer device Mar 2 14:25:40.791587 kernel: pstore: Using crash dump compression: deflate Mar 2 14:25:40.791598 kernel: pstore: Registered efi_pstore as persistent store backend Mar 2 14:25:40.791611 kernel: NET: Registered PF_INET6 protocol family Mar 2 14:25:40.791623 kernel: Segment Routing with IPv6 Mar 2 14:25:40.791634 kernel: In-situ OAM (IOAM) with IPv6 Mar 2 14:25:40.791646 kernel: NET: Registered PF_PACKET protocol family Mar 2 14:25:40.791662 kernel: Key type dns_resolver registered Mar 2 14:25:40.791674 kernel: IPI shorthand broadcast: enabled Mar 2 14:25:40.791686 kernel: sched_clock: Marking stable (7480020921, 804853913)->(9251995502, -967120668) Mar 2 14:25:40.791698 kernel: registered taskstats version 1 Mar 2 14:25:40.791711 kernel: Loading compiled-in X.509 certificates Mar 2 14:25:40.791724 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.74-flatcar: ca052fea375a75b056ebd4154b64794dffb70b96' Mar 2 14:25:40.791737 kernel: Demotion targets for Node 0: null Mar 2 14:25:40.791750 kernel: Key type .fscrypt registered Mar 2 14:25:40.791764 kernel: Key type fscrypt-provisioning registered Mar 2 14:25:40.791781 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 2 14:25:40.791793 kernel: ima: Allocated hash algorithm: sha1 Mar 2 14:25:40.791806 kernel: ima: No architecture policies found Mar 2 14:25:40.791820 kernel: clk: Disabling unused clocks Mar 2 14:25:40.791833 kernel: Warning: unable to open an initial console. Mar 2 14:25:40.791847 kernel: Freeing unused kernel image (initmem) memory: 46192K Mar 2 14:25:40.791861 kernel: Write protecting the kernel read-only data: 40960k Mar 2 14:25:40.791874 kernel: Freeing unused kernel image (rodata/data gap) memory: 560K Mar 2 14:25:40.791893 kernel: Run /init as init process Mar 2 14:25:40.791906 kernel: with arguments: Mar 2 14:25:40.791919 kernel: /init Mar 2 14:25:40.791932 kernel: with environment: Mar 2 14:25:40.791945 kernel: HOME=/ Mar 2 14:25:40.791959 kernel: TERM=linux Mar 2 14:25:40.791974 systemd[1]: Successfully made /usr/ read-only. Mar 2 14:25:40.791992 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 2 14:25:40.792011 systemd[1]: Detected virtualization kvm. Mar 2 14:25:40.792025 systemd[1]: Detected architecture x86-64. Mar 2 14:25:40.792038 systemd[1]: Running in initrd. Mar 2 14:25:40.792052 systemd[1]: No hostname configured, using default hostname. Mar 2 14:25:40.792066 systemd[1]: Hostname set to . Mar 2 14:25:40.792080 systemd[1]: Initializing machine ID from VM UUID. Mar 2 14:25:40.792094 systemd[1]: Queued start job for default target initrd.target. Mar 2 14:25:40.792108 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 14:25:40.792126 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 14:25:40.792141 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 2 14:25:40.792156 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 14:25:40.792170 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 2 14:25:40.792186 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 2 14:25:40.792201 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 2 14:25:40.792215 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 2 14:25:40.792303 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 14:25:40.792319 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 14:25:40.792332 systemd[1]: Reached target paths.target - Path Units. Mar 2 14:25:40.792346 systemd[1]: Reached target slices.target - Slice Units. Mar 2 14:25:40.792360 systemd[1]: Reached target swap.target - Swaps. Mar 2 14:25:40.792372 systemd[1]: Reached target timers.target - Timer Units. Mar 2 14:25:40.792385 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 14:25:40.792397 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 14:25:40.792414 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 2 14:25:40.792426 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Mar 2 14:25:40.792439 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 14:25:40.792451 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 14:25:40.792463 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 14:25:40.792476 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 14:25:40.792489 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 2 14:25:40.792502 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 14:25:40.792515 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 2 14:25:40.792533 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Mar 2 14:25:40.792546 systemd[1]: Starting systemd-fsck-usr.service... Mar 2 14:25:40.792558 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 14:25:40.792571 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 14:25:40.792584 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 14:25:40.792596 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 2 14:25:40.792613 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 14:25:40.792666 systemd-journald[204]: Collecting audit messages is disabled. Mar 2 14:25:40.792701 systemd[1]: Finished systemd-fsck-usr.service. Mar 2 14:25:40.792716 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 14:25:40.792731 systemd-journald[204]: Journal started Mar 2 14:25:40.792759 systemd-journald[204]: Runtime Journal (/run/log/journal/402c96098cf44df7b40f02fd7ebec8e6) is 6M, max 48.1M, 42.1M free. Mar 2 14:25:40.785480 systemd-modules-load[206]: Inserted module 'overlay' Mar 2 14:25:40.820537 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 14:25:40.833342 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 14:25:40.837999 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 2 14:25:40.855735 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 14:25:40.889336 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 2 14:25:40.891575 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 14:25:40.900474 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 14:25:40.919685 kernel: Bridge firewalling registered Mar 2 14:25:40.918641 systemd-modules-load[206]: Inserted module 'br_netfilter' Mar 2 14:25:40.926693 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 14:25:40.928174 systemd-tmpfiles[220]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Mar 2 14:25:40.945084 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 14:25:40.951987 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 14:25:40.973957 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 14:25:40.987526 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 14:25:41.001627 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 14:25:41.008080 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 2 14:25:41.018104 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 14:25:41.079653 dracut-cmdline[245]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=82731586f036a8515942386c762f58de23efa7b4e7ecf4198e267e112154cbc2 Mar 2 14:25:41.142784 systemd-resolved[246]: Positive Trust Anchors: Mar 2 14:25:41.142825 systemd-resolved[246]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 14:25:41.142870 systemd-resolved[246]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 14:25:41.149949 systemd-resolved[246]: Defaulting to hostname 'linux'. Mar 2 14:25:41.154559 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 14:25:41.157851 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 14:25:41.392493 kernel: SCSI subsystem initialized Mar 2 14:25:41.408370 kernel: Loading iSCSI transport class v2.0-870. Mar 2 14:25:41.436138 kernel: iscsi: registered transport (tcp) Mar 2 14:25:41.474566 kernel: iscsi: registered transport (qla4xxx) Mar 2 14:25:41.474661 kernel: QLogic iSCSI HBA Driver Mar 2 14:25:41.515604 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 14:25:41.560884 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 14:25:41.570933 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 14:25:41.700150 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 2 14:25:41.708683 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 2 14:25:41.844494 kernel: raid6: avx2x4 gen() 19675 MB/s Mar 2 14:25:41.863505 kernel: raid6: avx2x2 gen() 17238 MB/s Mar 2 14:25:41.884111 kernel: raid6: avx2x1 gen() 11345 MB/s Mar 2 14:25:41.884370 kernel: raid6: using algorithm avx2x4 gen() 19675 MB/s Mar 2 14:25:41.906436 kernel: raid6: .... xor() 5682 MB/s, rmw enabled Mar 2 14:25:41.906512 kernel: raid6: using avx2x2 recovery algorithm Mar 2 14:25:41.949420 kernel: xor: automatically using best checksumming function avx Mar 2 14:25:42.479839 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 2 14:25:42.500162 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 2 14:25:42.514811 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 14:25:42.577589 systemd-udevd[456]: Using default interface naming scheme 'v255'. Mar 2 14:25:42.591560 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 14:25:42.595731 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 2 14:25:42.664595 dracut-pre-trigger[458]: rd.md=0: removing MD RAID activation Mar 2 14:25:42.734394 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 14:25:42.739569 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 14:25:42.977707 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 14:25:42.982679 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 2 14:25:43.094538 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Mar 2 14:25:43.114042 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Mar 2 14:25:43.119663 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 2 14:25:43.126460 kernel: GPT:9289727 != 19775487 Mar 2 14:25:43.126504 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 2 14:25:43.132061 kernel: GPT:9289727 != 19775487 Mar 2 14:25:43.132099 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 2 14:25:43.140695 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 14:25:43.179894 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 14:25:43.198731 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Mar 2 14:25:43.180123 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 14:25:43.210621 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 14:25:43.238665 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 14:25:43.244352 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 2 14:25:43.260389 kernel: libata version 3.00 loaded. Mar 2 14:25:43.260435 kernel: cryptd: max_cpu_qlen set to 1000 Mar 2 14:25:43.280316 kernel: ahci 0000:00:1f.2: version 3.0 Mar 2 14:25:43.280584 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 2 14:25:43.287902 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 2 14:25:43.299135 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Mar 2 14:25:43.299444 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Mar 2 14:25:43.299671 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 2 14:25:43.380649 kernel: AES CTR mode by8 optimization enabled Mar 2 14:25:43.380743 kernel: scsi host0: ahci Mar 2 14:25:43.386608 kernel: scsi host1: ahci Mar 2 14:25:43.389714 kernel: scsi host2: ahci Mar 2 14:25:43.400358 kernel: scsi host3: ahci Mar 2 14:25:43.407173 kernel: scsi host4: ahci Mar 2 14:25:43.409349 kernel: scsi host5: ahci Mar 2 14:25:43.414961 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 2 14:25:43.462665 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 lpm-pol 1 Mar 2 14:25:43.462701 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 lpm-pol 1 Mar 2 14:25:43.462719 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 lpm-pol 1 Mar 2 14:25:43.462736 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 lpm-pol 1 Mar 2 14:25:43.462748 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 lpm-pol 1 Mar 2 14:25:43.462761 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 lpm-pol 1 Mar 2 14:25:43.470842 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 2 14:25:43.499778 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Mar 2 14:25:43.501371 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 2 14:25:43.534907 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 2 14:25:43.545029 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 14:25:43.545686 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 14:25:43.566949 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 14:25:43.603629 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 14:25:43.610146 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 2 14:25:43.644802 disk-uuid[616]: Primary Header is updated. Mar 2 14:25:43.644802 disk-uuid[616]: Secondary Entries is updated. Mar 2 14:25:43.644802 disk-uuid[616]: Secondary Header is updated. Mar 2 14:25:43.660458 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 14:25:43.675895 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 14:25:43.704395 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 14:25:43.738750 kernel: ata1: SATA link down (SStatus 0 SControl 300) Mar 2 14:25:43.738808 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 2 14:25:43.739849 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 2 14:25:43.749328 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Mar 2 14:25:43.749464 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 2 14:25:43.752319 kernel: ata3.00: LPM support broken, forcing max_power Mar 2 14:25:43.761727 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Mar 2 14:25:43.761805 kernel: ata3.00: applying bridge limits Mar 2 14:25:43.767052 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 2 14:25:43.776927 kernel: ata3.00: LPM support broken, forcing max_power Mar 2 14:25:43.776962 kernel: ata3.00: configured for UDMA/100 Mar 2 14:25:43.803046 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 2 14:25:43.937212 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Mar 2 14:25:43.938035 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 2 14:25:43.959152 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Mar 2 14:25:44.451619 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 2 14:25:44.454706 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 14:25:44.475051 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 14:25:44.483456 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 14:25:44.486364 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 2 14:25:44.552659 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 2 14:25:44.684152 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 14:25:44.686411 disk-uuid[618]: The operation has completed successfully. Mar 2 14:25:44.772991 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 2 14:25:44.773187 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 2 14:25:44.831004 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 2 14:25:44.870769 sh[662]: Success Mar 2 14:25:44.928606 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 2 14:25:44.928682 kernel: device-mapper: uevent: version 1.0.3 Mar 2 14:25:44.933381 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Mar 2 14:25:44.965365 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Mar 2 14:25:45.074166 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 2 14:25:45.090456 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 2 14:25:45.105083 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 2 14:25:45.142414 kernel: BTRFS: device fsid 760529e6-8e55-47fc-ad5a-c1c1d184e50a devid 1 transid 37 /dev/mapper/usr (253:0) scanned by mount (674) Mar 2 14:25:45.152347 kernel: BTRFS info (device dm-0): first mount of filesystem 760529e6-8e55-47fc-ad5a-c1c1d184e50a Mar 2 14:25:45.152410 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 2 14:25:45.189224 kernel: BTRFS info (device dm-0 state E): disabling log replay at mount time Mar 2 14:25:45.189372 kernel: BTRFS info (device dm-0 state E): enabling free space tree Mar 2 14:25:45.196356 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 2 14:25:45.197463 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Mar 2 14:25:45.217310 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 2 14:25:45.220953 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 2 14:25:45.243919 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 2 14:25:45.347366 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (706) Mar 2 14:25:45.359712 kernel: BTRFS info (device vda6): first mount of filesystem 81b29f52-362f-4f57-bc73-813781f2dfeb Mar 2 14:25:45.359778 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 14:25:45.372086 kernel: BTRFS info (device vda6): turning on async discard Mar 2 14:25:45.372163 kernel: BTRFS info (device vda6): enabling free space tree Mar 2 14:25:45.385577 kernel: BTRFS info (device vda6): last unmount of filesystem 81b29f52-362f-4f57-bc73-813781f2dfeb Mar 2 14:25:45.393568 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 2 14:25:45.397575 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 2 14:25:45.629816 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 14:25:45.737702 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 14:25:45.907491 systemd-networkd[843]: lo: Link UP Mar 2 14:25:45.907561 systemd-networkd[843]: lo: Gained carrier Mar 2 14:25:45.942405 systemd-networkd[843]: Enumeration completed Mar 2 14:25:45.943049 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 14:25:45.943462 systemd-networkd[843]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 14:25:45.943468 systemd-networkd[843]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 14:25:45.952450 systemd-networkd[843]: eth0: Link UP Mar 2 14:25:45.953453 systemd-networkd[843]: eth0: Gained carrier Mar 2 14:25:45.953469 systemd-networkd[843]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 14:25:45.962018 systemd[1]: Reached target network.target - Network. Mar 2 14:25:46.008141 systemd-networkd[843]: eth0: DHCPv4 address 10.0.0.9/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 2 14:25:46.244737 ignition[763]: Ignition 2.22.0 Mar 2 14:25:46.244751 ignition[763]: Stage: fetch-offline Mar 2 14:25:46.244864 ignition[763]: no configs at "/usr/lib/ignition/base.d" Mar 2 14:25:46.244879 ignition[763]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 14:25:46.245098 ignition[763]: parsed url from cmdline: "" Mar 2 14:25:46.245107 ignition[763]: no config URL provided Mar 2 14:25:46.245117 ignition[763]: reading system config file "/usr/lib/ignition/user.ign" Mar 2 14:25:46.245130 ignition[763]: no config at "/usr/lib/ignition/user.ign" Mar 2 14:25:46.245509 ignition[763]: op(1): [started] loading QEMU firmware config module Mar 2 14:25:46.245518 ignition[763]: op(1): executing: "modprobe" "qemu_fw_cfg" Mar 2 14:25:46.318358 ignition[763]: op(1): [finished] loading QEMU firmware config module Mar 2 14:25:46.574599 ignition[763]: parsing config with SHA512: 418d7dfb796be322ee1b5b024e3a61792767ee352ff4dcb3900f3d6bc2e363c667f04d2f8b7ebd1f7fb2e8c981f42857f14336c733417d12ce7f66d3406ad558 Mar 2 14:25:46.586990 unknown[763]: fetched base config from "system" Mar 2 14:25:46.587006 unknown[763]: fetched user config from "qemu" Mar 2 14:25:46.587847 ignition[763]: fetch-offline: fetch-offline passed Mar 2 14:25:46.587954 ignition[763]: Ignition finished successfully Mar 2 14:25:46.604017 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 14:25:46.608511 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Mar 2 14:25:46.611142 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 2 14:25:46.741169 ignition[856]: Ignition 2.22.0 Mar 2 14:25:46.751026 ignition[856]: Stage: kargs Mar 2 14:25:46.755365 ignition[856]: no configs at "/usr/lib/ignition/base.d" Mar 2 14:25:46.755456 ignition[856]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 14:25:46.764813 ignition[856]: kargs: kargs passed Mar 2 14:25:46.764930 ignition[856]: Ignition finished successfully Mar 2 14:25:46.804530 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 2 14:25:46.820433 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 2 14:25:47.036222 ignition[864]: Ignition 2.22.0 Mar 2 14:25:47.036387 ignition[864]: Stage: disks Mar 2 14:25:47.036575 ignition[864]: no configs at "/usr/lib/ignition/base.d" Mar 2 14:25:47.043822 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 2 14:25:47.036587 ignition[864]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 14:25:47.049861 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 2 14:25:47.038085 ignition[864]: disks: disks passed Mar 2 14:25:47.060585 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 2 14:25:47.038147 ignition[864]: Ignition finished successfully Mar 2 14:25:47.064634 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 14:25:47.078859 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 14:25:47.087930 systemd[1]: Reached target basic.target - Basic System. Mar 2 14:25:47.110936 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 2 14:25:47.291134 systemd-fsck[874]: ROOT: clean, 15/553520 files, 52789/553472 blocks Mar 2 14:25:47.323169 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 2 14:25:47.341434 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 2 14:25:47.696489 systemd-networkd[843]: eth0: Gained IPv6LL Mar 2 14:25:48.040454 kernel: EXT4-fs (vda9): mounted filesystem 9d55f1a4-66ad-43d6-b325-f6b8d2d08c3e r/w with ordered data mode. Quota mode: none. Mar 2 14:25:48.045501 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 2 14:25:48.062936 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 2 14:25:48.076583 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 14:25:48.089860 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 2 14:25:48.095489 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 2 14:25:48.095567 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 2 14:25:48.095608 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 14:25:48.163528 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 2 14:25:48.179502 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 2 14:25:48.197322 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (882) Mar 2 14:25:48.222362 kernel: BTRFS info (device vda6): first mount of filesystem 81b29f52-362f-4f57-bc73-813781f2dfeb Mar 2 14:25:48.222438 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 14:25:48.251967 kernel: BTRFS info (device vda6): turning on async discard Mar 2 14:25:48.252043 kernel: BTRFS info (device vda6): enabling free space tree Mar 2 14:25:48.257359 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 14:25:48.372605 initrd-setup-root[906]: cut: /sysroot/etc/passwd: No such file or directory Mar 2 14:25:48.389531 initrd-setup-root[913]: cut: /sysroot/etc/group: No such file or directory Mar 2 14:25:48.411748 initrd-setup-root[920]: cut: /sysroot/etc/shadow: No such file or directory Mar 2 14:25:48.452463 initrd-setup-root[927]: cut: /sysroot/etc/gshadow: No such file or directory Mar 2 14:25:49.030941 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 2 14:25:49.050045 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 2 14:25:49.088178 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 2 14:25:49.138121 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 2 14:25:49.147428 kernel: BTRFS info (device vda6): last unmount of filesystem 81b29f52-362f-4f57-bc73-813781f2dfeb Mar 2 14:25:49.221864 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 2 14:25:49.301110 ignition[994]: INFO : Ignition 2.22.0 Mar 2 14:25:49.305792 ignition[994]: INFO : Stage: mount Mar 2 14:25:49.305792 ignition[994]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 14:25:49.305792 ignition[994]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 14:25:49.305792 ignition[994]: INFO : mount: mount passed Mar 2 14:25:49.305792 ignition[994]: INFO : Ignition finished successfully Mar 2 14:25:49.317184 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 2 14:25:49.329198 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 2 14:25:49.471559 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 14:25:49.624512 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1007) Mar 2 14:25:49.643997 kernel: BTRFS info (device vda6): first mount of filesystem 81b29f52-362f-4f57-bc73-813781f2dfeb Mar 2 14:25:49.644328 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 14:25:49.667083 kernel: BTRFS info (device vda6): turning on async discard Mar 2 14:25:49.667170 kernel: BTRFS info (device vda6): enabling free space tree Mar 2 14:25:49.677471 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 14:25:49.808815 ignition[1023]: INFO : Ignition 2.22.0 Mar 2 14:25:49.808815 ignition[1023]: INFO : Stage: files Mar 2 14:25:49.808815 ignition[1023]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 14:25:49.808815 ignition[1023]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 14:25:49.851856 ignition[1023]: DEBUG : files: compiled without relabeling support, skipping Mar 2 14:25:49.851856 ignition[1023]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 2 14:25:49.851856 ignition[1023]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 2 14:25:49.873756 ignition[1023]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 2 14:25:49.873756 ignition[1023]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 2 14:25:49.873756 ignition[1023]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 2 14:25:49.868402 unknown[1023]: wrote ssh authorized keys file for user: core Mar 2 14:25:49.902705 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 2 14:25:49.902705 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Mar 2 14:25:50.021195 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 2 14:25:50.810524 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 2 14:25:50.831222 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Mar 2 14:25:50.831222 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Mar 2 14:25:50.831222 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 2 14:25:50.831222 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 2 14:25:50.831222 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 14:25:50.831222 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 14:25:50.831222 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 14:25:50.831222 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 14:25:50.971838 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 14:25:50.971838 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 14:25:50.971838 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 2 14:25:50.971838 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 2 14:25:50.971838 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 2 14:25:50.971838 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.34.4-x86-64.raw: attempt #1 Mar 2 14:25:51.425411 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Mar 2 14:25:52.470813 kernel: hrtimer: interrupt took 3856371 ns Mar 2 14:25:52.625955 ignition[1023]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 2 14:25:52.625955 ignition[1023]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Mar 2 14:25:52.642759 ignition[1023]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 14:25:52.652966 ignition[1023]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 14:25:52.652966 ignition[1023]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Mar 2 14:25:52.652966 ignition[1023]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Mar 2 14:25:52.652966 ignition[1023]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 2 14:25:52.652966 ignition[1023]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 2 14:25:52.652966 ignition[1023]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Mar 2 14:25:52.652966 ignition[1023]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Mar 2 14:25:52.737559 ignition[1023]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Mar 2 14:25:52.748127 ignition[1023]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Mar 2 14:25:52.748127 ignition[1023]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Mar 2 14:25:52.764796 ignition[1023]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Mar 2 14:25:52.764796 ignition[1023]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Mar 2 14:25:52.764796 ignition[1023]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 2 14:25:52.764796 ignition[1023]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 2 14:25:52.764796 ignition[1023]: INFO : files: files passed Mar 2 14:25:52.764796 ignition[1023]: INFO : Ignition finished successfully Mar 2 14:25:52.773498 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 2 14:25:52.808209 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 2 14:25:52.818525 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 2 14:25:52.886484 initrd-setup-root-after-ignition[1052]: grep: /sysroot/oem/oem-release: No such file or directory Mar 2 14:25:52.904101 initrd-setup-root-after-ignition[1055]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 14:25:52.904101 initrd-setup-root-after-ignition[1055]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 2 14:25:52.890980 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 2 14:25:52.936593 initrd-setup-root-after-ignition[1059]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 14:25:52.891151 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 2 14:25:52.913693 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 14:25:52.934065 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 2 14:25:52.968028 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 2 14:25:53.072845 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 2 14:25:53.073094 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 2 14:25:53.075620 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 2 14:25:53.075954 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 2 14:25:53.108200 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 2 14:25:53.114528 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 2 14:25:53.143828 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 14:25:53.157832 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 2 14:25:53.204143 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 2 14:25:53.212080 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 14:25:53.216971 systemd[1]: Stopped target timers.target - Timer Units. Mar 2 14:25:53.227636 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 2 14:25:53.227872 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 14:25:53.242614 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 2 14:25:53.245645 systemd[1]: Stopped target basic.target - Basic System. Mar 2 14:25:53.271401 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 2 14:25:53.281190 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 14:25:53.281459 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 2 14:25:53.298769 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Mar 2 14:25:53.311209 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 2 14:25:53.325445 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 14:25:53.345697 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 2 14:25:53.367520 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 2 14:25:53.377983 systemd[1]: Stopped target swap.target - Swaps. Mar 2 14:25:53.387627 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 2 14:25:53.387812 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 2 14:25:53.399804 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 2 14:25:53.409908 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 14:25:53.416777 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 2 14:25:53.424966 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 14:25:53.435448 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 2 14:25:53.435678 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 2 14:25:53.442992 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 2 14:25:53.444109 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 14:25:53.456817 systemd[1]: Stopped target paths.target - Path Units. Mar 2 14:25:53.459810 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 2 14:25:53.472031 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 14:25:53.479069 systemd[1]: Stopped target slices.target - Slice Units. Mar 2 14:25:53.479342 systemd[1]: Stopped target sockets.target - Socket Units. Mar 2 14:25:53.480015 systemd[1]: iscsid.socket: Deactivated successfully. Mar 2 14:25:53.480163 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 14:25:53.547882 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 2 14:25:53.548111 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 14:25:53.563756 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 2 14:25:53.566750 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 14:25:53.571574 systemd[1]: ignition-files.service: Deactivated successfully. Mar 2 14:25:53.571730 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 2 14:25:53.587158 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 2 14:25:53.633656 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 2 14:25:53.640503 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 2 14:25:53.640711 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 14:25:53.655023 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 2 14:25:53.656327 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 14:25:53.681532 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 2 14:25:53.695122 ignition[1080]: INFO : Ignition 2.22.0 Mar 2 14:25:53.695122 ignition[1080]: INFO : Stage: umount Mar 2 14:25:53.695122 ignition[1080]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 14:25:53.695122 ignition[1080]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 14:25:53.681718 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 2 14:25:53.710762 ignition[1080]: INFO : umount: umount passed Mar 2 14:25:53.710762 ignition[1080]: INFO : Ignition finished successfully Mar 2 14:25:53.701114 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 2 14:25:53.701428 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 2 14:25:53.735090 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 2 14:25:53.744069 systemd[1]: Stopped target network.target - Network. Mar 2 14:25:53.744989 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 2 14:25:53.745102 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 2 14:25:53.754498 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 2 14:25:53.754607 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 2 14:25:53.755167 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 2 14:25:53.755332 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 2 14:25:53.755769 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 2 14:25:53.755841 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 2 14:25:53.763027 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 2 14:25:53.824973 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 2 14:25:53.839643 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 2 14:25:53.839808 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 2 14:25:53.859103 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Mar 2 14:25:53.860764 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 2 14:25:53.860972 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 2 14:25:53.888602 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Mar 2 14:25:53.889071 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 2 14:25:53.890146 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 2 14:25:53.906996 systemd[1]: Stopped target network-pre.target - Preparation for Network. Mar 2 14:25:53.913373 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 2 14:25:53.913462 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 2 14:25:53.913893 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 2 14:25:53.913973 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 2 14:25:53.922501 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 2 14:25:53.938023 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 2 14:25:53.938132 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 14:25:53.938354 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 2 14:25:53.938424 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 2 14:25:53.957887 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 2 14:25:53.957994 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 2 14:25:53.960947 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 2 14:25:53.961025 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 14:25:54.000722 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 14:25:54.008019 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 2 14:25:54.008129 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Mar 2 14:25:54.057556 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 2 14:25:54.066711 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 14:25:54.121388 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 2 14:25:54.121534 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 2 14:25:54.128207 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 2 14:25:54.131278 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 14:25:54.148224 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 2 14:25:54.148490 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 2 14:25:54.168568 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 2 14:25:54.168684 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 2 14:25:54.181800 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 2 14:25:54.182066 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 14:25:54.198489 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 2 14:25:54.227873 systemd[1]: systemd-network-generator.service: Deactivated successfully. Mar 2 14:25:54.228033 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 14:25:54.246640 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 2 14:25:54.246801 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 14:25:54.261975 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Mar 2 14:25:54.262130 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 14:25:54.279075 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 2 14:25:54.279183 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 14:25:54.279959 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 14:25:54.280016 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 14:25:54.312578 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Mar 2 14:25:54.312708 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Mar 2 14:25:54.312780 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Mar 2 14:25:54.312856 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 2 14:25:54.313979 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 2 14:25:54.314344 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 2 14:25:54.324737 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 2 14:25:54.324938 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 2 14:25:54.336003 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 2 14:25:54.353807 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 2 14:25:54.432800 systemd[1]: Switching root. Mar 2 14:25:54.482726 systemd-journald[204]: Journal stopped Mar 2 14:25:57.935147 systemd-journald[204]: Received SIGTERM from PID 1 (systemd). Mar 2 14:25:57.935393 kernel: SELinux: policy capability network_peer_controls=1 Mar 2 14:25:57.935414 kernel: SELinux: policy capability open_perms=1 Mar 2 14:25:57.935430 kernel: SELinux: policy capability extended_socket_class=1 Mar 2 14:25:57.935444 kernel: SELinux: policy capability always_check_network=0 Mar 2 14:25:57.935468 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 2 14:25:57.935484 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 2 14:25:57.935715 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 2 14:25:57.935736 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 2 14:25:57.935751 kernel: SELinux: policy capability userspace_initial_context=0 Mar 2 14:25:57.935771 kernel: audit: type=1403 audit(1772461554.918:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 2 14:25:57.935798 systemd[1]: Successfully loaded SELinux policy in 120.989ms. Mar 2 14:25:57.935913 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 14.861ms. Mar 2 14:25:57.935936 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 2 14:25:57.935955 systemd[1]: Detected virtualization kvm. Mar 2 14:25:57.935971 systemd[1]: Detected architecture x86-64. Mar 2 14:25:57.935986 systemd[1]: Detected first boot. Mar 2 14:25:57.936001 systemd[1]: Initializing machine ID from VM UUID. Mar 2 14:25:57.936017 zram_generator::config[1128]: No configuration found. Mar 2 14:25:57.936041 kernel: Guest personality initialized and is inactive Mar 2 14:25:57.936056 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Mar 2 14:25:57.936070 kernel: Initialized host personality Mar 2 14:25:57.936084 kernel: NET: Registered PF_VSOCK protocol family Mar 2 14:25:57.936099 systemd[1]: Populated /etc with preset unit settings. Mar 2 14:25:57.936123 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Mar 2 14:25:57.936138 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 2 14:25:57.936161 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 2 14:25:57.936181 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 2 14:25:57.936199 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 2 14:25:57.936219 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 2 14:25:57.936312 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 2 14:25:57.936370 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 2 14:25:57.936392 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 2 14:25:57.936409 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 2 14:25:57.936424 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 2 14:25:57.936439 systemd[1]: Created slice user.slice - User and Session Slice. Mar 2 14:25:57.936461 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 14:25:57.936482 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 14:25:57.936501 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 2 14:25:57.936516 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 2 14:25:57.936532 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 2 14:25:57.936548 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 14:25:57.936568 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 2 14:25:57.936589 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 14:25:57.936604 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 14:25:57.937413 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 2 14:25:57.937433 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 2 14:25:57.937448 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 2 14:25:57.937501 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 2 14:25:57.937517 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 14:25:57.937533 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 14:25:57.937548 systemd[1]: Reached target slices.target - Slice Units. Mar 2 14:25:57.937563 systemd[1]: Reached target swap.target - Swaps. Mar 2 14:25:57.937586 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 2 14:25:57.937601 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 2 14:25:57.937615 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Mar 2 14:25:57.937631 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 14:25:57.937647 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 14:25:57.937662 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 14:25:57.937678 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 2 14:25:57.937693 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 2 14:25:57.937708 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 2 14:25:57.937928 systemd[1]: Mounting media.mount - External Media Directory... Mar 2 14:25:57.937946 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 14:25:57.937962 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 2 14:25:57.937977 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 2 14:25:57.937992 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 2 14:25:57.938008 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 2 14:25:57.938023 systemd[1]: Reached target machines.target - Containers. Mar 2 14:25:57.938070 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 2 14:25:57.938090 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 14:25:57.938105 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 14:25:57.938120 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 2 14:25:57.938136 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 14:25:57.938151 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 14:25:57.938166 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 14:25:57.938181 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 2 14:25:57.938200 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 14:25:57.938220 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 2 14:25:57.938291 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 2 14:25:57.938310 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 2 14:25:57.938362 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 2 14:25:57.938383 systemd[1]: Stopped systemd-fsck-usr.service. Mar 2 14:25:57.938398 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 14:25:57.938413 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 14:25:57.938428 kernel: loop: module loaded Mar 2 14:25:57.938443 kernel: fuse: init (API version 7.41) Mar 2 14:25:57.938463 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 14:25:57.938479 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 14:25:57.938497 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 2 14:25:57.938587 systemd-journald[1213]: Collecting audit messages is disabled. Mar 2 14:25:57.938623 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Mar 2 14:25:57.938639 systemd-journald[1213]: Journal started Mar 2 14:25:57.938667 systemd-journald[1213]: Runtime Journal (/run/log/journal/402c96098cf44df7b40f02fd7ebec8e6) is 6M, max 48.1M, 42.1M free. Mar 2 14:25:56.454487 systemd[1]: Queued start job for default target multi-user.target. Mar 2 14:25:56.498999 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 2 14:25:56.513460 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 2 14:25:56.517291 systemd[1]: systemd-journald.service: Consumed 1.224s CPU time. Mar 2 14:25:57.968009 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 14:25:57.981792 systemd[1]: verity-setup.service: Deactivated successfully. Mar 2 14:25:57.982433 systemd[1]: Stopped verity-setup.service. Mar 2 14:25:57.993373 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 14:25:57.997441 kernel: ACPI: bus type drm_connector registered Mar 2 14:25:58.008505 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 14:25:58.023662 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 2 14:25:58.027673 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 2 14:25:58.034093 systemd[1]: Mounted media.mount - External Media Directory. Mar 2 14:25:58.039998 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 2 14:25:58.044921 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 2 14:25:58.049632 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 2 14:25:58.200059 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 2 14:25:58.223850 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 14:25:58.232374 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 2 14:25:58.233497 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 2 14:25:58.242400 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 14:25:58.243059 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 14:25:58.249693 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 14:25:58.250048 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 14:25:58.259822 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 14:25:58.260303 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 14:25:58.269737 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 2 14:25:58.270105 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 2 14:25:58.278115 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 14:25:58.278754 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 14:25:58.293071 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 14:25:58.303738 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 14:25:58.311620 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 2 14:25:58.329778 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Mar 2 14:25:58.355710 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 14:25:58.376926 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 14:25:58.387557 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 2 14:25:58.441026 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 2 14:25:58.449215 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 2 14:25:58.449383 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 14:25:58.459172 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Mar 2 14:25:58.468619 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 2 14:25:58.477478 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 14:25:58.490009 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 2 14:25:58.501606 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 2 14:25:58.506385 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 14:25:58.514802 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 2 14:25:58.525350 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 14:25:58.541726 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 14:25:58.563094 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 2 14:25:58.585502 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 14:25:58.596619 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 2 14:25:58.605910 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 2 14:25:58.606723 systemd-journald[1213]: Time spent on flushing to /var/log/journal/402c96098cf44df7b40f02fd7ebec8e6 is 150.630ms for 1072 entries. Mar 2 14:25:58.606723 systemd-journald[1213]: System Journal (/var/log/journal/402c96098cf44df7b40f02fd7ebec8e6) is 8M, max 195.6M, 187.6M free. Mar 2 14:25:58.828375 systemd-journald[1213]: Received client request to flush runtime journal. Mar 2 14:25:58.828458 kernel: loop0: detected capacity change from 0 to 110984 Mar 2 14:25:58.629203 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 2 14:25:58.650602 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 2 14:25:58.660510 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Mar 2 14:25:58.693597 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 14:25:58.780931 systemd-tmpfiles[1249]: ACLs are not supported, ignoring. Mar 2 14:25:58.780952 systemd-tmpfiles[1249]: ACLs are not supported, ignoring. Mar 2 14:25:58.823960 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 14:25:58.831515 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 2 14:25:58.846783 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 2 14:25:58.846915 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 2 14:25:58.849166 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Mar 2 14:25:58.866473 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 2 14:25:58.886309 kernel: loop1: detected capacity change from 0 to 128560 Mar 2 14:25:58.964580 kernel: loop2: detected capacity change from 0 to 219192 Mar 2 14:25:58.975937 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 2 14:25:58.986406 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 14:25:59.625584 kernel: loop3: detected capacity change from 0 to 110984 Mar 2 14:25:59.629220 systemd-tmpfiles[1273]: ACLs are not supported, ignoring. Mar 2 14:25:59.629313 systemd-tmpfiles[1273]: ACLs are not supported, ignoring. Mar 2 14:25:59.640172 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 14:25:59.674420 kernel: loop4: detected capacity change from 0 to 128560 Mar 2 14:25:59.755316 kernel: loop5: detected capacity change from 0 to 219192 Mar 2 14:25:59.831922 (sd-merge)[1276]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Mar 2 14:25:59.832913 (sd-merge)[1276]: Merged extensions into '/usr'. Mar 2 14:25:59.856406 systemd[1]: Reload requested from client PID 1248 ('systemd-sysext') (unit systemd-sysext.service)... Mar 2 14:25:59.856430 systemd[1]: Reloading... Mar 2 14:26:00.056404 zram_generator::config[1306]: No configuration found. Mar 2 14:26:00.218359 ldconfig[1243]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 2 14:26:00.408536 systemd[1]: Reloading finished in 546 ms. Mar 2 14:26:00.434499 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 2 14:26:00.440783 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 2 14:26:00.446173 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 2 14:26:00.473869 systemd[1]: Starting ensure-sysext.service... Mar 2 14:26:00.487536 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 14:26:00.516885 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 14:26:00.559498 systemd-tmpfiles[1342]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Mar 2 14:26:00.559583 systemd-tmpfiles[1342]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Mar 2 14:26:00.560030 systemd-tmpfiles[1342]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 2 14:26:00.561638 systemd-tmpfiles[1342]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 2 14:26:00.566129 systemd-tmpfiles[1342]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 2 14:26:00.570018 systemd-tmpfiles[1342]: ACLs are not supported, ignoring. Mar 2 14:26:00.573704 systemd-tmpfiles[1342]: ACLs are not supported, ignoring. Mar 2 14:26:00.588025 systemd-tmpfiles[1342]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 14:26:00.588033 systemd[1]: Reload requested from client PID 1341 ('systemctl') (unit ensure-sysext.service)... Mar 2 14:26:00.588061 systemd-tmpfiles[1342]: Skipping /boot Mar 2 14:26:00.588062 systemd[1]: Reloading... Mar 2 14:26:00.605161 systemd-tmpfiles[1342]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 14:26:00.605178 systemd-tmpfiles[1342]: Skipping /boot Mar 2 14:26:00.641587 systemd-udevd[1343]: Using default interface naming scheme 'v255'. Mar 2 14:26:00.694851 zram_generator::config[1373]: No configuration found. Mar 2 14:26:00.946367 kernel: mousedev: PS/2 mouse device common for all mice Mar 2 14:26:00.981506 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Mar 2 14:26:00.988799 kernel: ACPI: button: Power Button [PWRF] Mar 2 14:26:01.016320 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Mar 2 14:26:01.028692 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 2 14:26:01.029069 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 2 14:26:01.042105 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Mar 2 14:26:01.042772 systemd[1]: Reloading finished in 453 ms. Mar 2 14:26:01.076669 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 14:26:01.100456 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 14:26:01.154820 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 2 14:26:01.165101 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 14:26:01.169829 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 2 14:26:01.182862 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 2 14:26:01.187693 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 14:26:01.193432 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 14:26:01.202085 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 14:26:01.208055 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 14:26:01.213699 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 14:26:01.215924 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 2 14:26:01.221520 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 14:26:01.236552 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 2 14:26:01.279474 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 14:26:01.298011 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 14:26:01.308304 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 2 14:26:01.309053 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 14:26:01.327987 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 14:26:01.329588 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 14:26:01.400746 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 2 14:26:01.450891 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 14:26:01.456981 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 14:26:01.466441 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 14:26:01.500119 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 14:26:01.536278 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 2 14:26:01.551073 systemd[1]: Finished ensure-sysext.service. Mar 2 14:26:01.557876 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 2 14:26:01.582765 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 14:26:01.583127 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 14:26:01.587316 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 14:26:01.599400 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 14:26:01.610607 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 14:26:01.634702 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 14:26:01.639626 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 14:26:01.639788 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 14:26:01.643855 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 2 14:26:01.653755 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 2 14:26:01.675828 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 2 14:26:01.681496 augenrules[1504]: No rules Mar 2 14:26:01.685780 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 14:26:01.690803 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 14:26:01.696079 systemd[1]: audit-rules.service: Deactivated successfully. Mar 2 14:26:01.696920 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 2 14:26:01.720292 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 2 14:26:01.744900 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 14:26:01.745334 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 14:26:01.757631 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 14:26:01.849599 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 14:26:01.853981 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 14:26:01.855647 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 14:26:01.862825 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 14:26:01.863138 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 14:26:01.866075 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 2 14:26:01.888137 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 14:26:01.888328 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 14:26:01.888404 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 2 14:26:01.965586 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 14:26:02.034440 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 2 14:26:02.228639 kernel: kvm_amd: TSC scaling supported Mar 2 14:26:02.228732 kernel: kvm_amd: Nested Virtualization enabled Mar 2 14:26:02.228789 kernel: kvm_amd: Nested Paging enabled Mar 2 14:26:02.235472 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Mar 2 14:26:02.235557 kernel: kvm_amd: PMU virtualization is disabled Mar 2 14:26:02.250187 systemd-resolved[1474]: Positive Trust Anchors: Mar 2 14:26:02.250704 systemd-resolved[1474]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 14:26:02.250755 systemd-resolved[1474]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 14:26:02.267584 systemd-resolved[1474]: Defaulting to hostname 'linux'. Mar 2 14:26:02.271825 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 14:26:02.278697 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 14:26:02.285406 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 2 14:26:02.290936 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 14:26:02.291775 systemd-networkd[1473]: lo: Link UP Mar 2 14:26:02.291782 systemd-networkd[1473]: lo: Gained carrier Mar 2 14:26:02.295488 systemd-networkd[1473]: Enumeration completed Mar 2 14:26:02.296579 systemd-networkd[1473]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 14:26:02.296611 systemd-networkd[1473]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 14:26:02.297124 systemd-networkd[1473]: eth0: Link UP Mar 2 14:26:02.300512 systemd-networkd[1473]: eth0: Gained carrier Mar 2 14:26:02.300560 systemd-networkd[1473]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 14:26:02.301935 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 2 14:26:02.316919 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 2 14:26:02.322120 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Mar 2 14:26:02.328307 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 2 14:26:02.334188 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 2 14:26:02.335199 systemd[1]: Reached target paths.target - Path Units. Mar 2 14:26:02.340211 systemd[1]: Reached target time-set.target - System Time Set. Mar 2 14:26:02.345588 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 2 14:26:02.351116 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 2 14:26:02.358751 systemd[1]: Reached target timers.target - Timer Units. Mar 2 14:26:02.368969 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 2 14:26:02.379717 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 2 14:26:02.392549 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Mar 2 14:26:02.403440 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Mar 2 14:26:02.415112 systemd[1]: Reached target ssh-access.target - SSH Access Available. Mar 2 14:26:02.449756 systemd-networkd[1473]: eth0: DHCPv4 address 10.0.0.9/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 2 14:26:02.453675 systemd-timesyncd[1499]: Network configuration changed, trying to establish connection. Mar 2 14:26:03.127059 systemd-resolved[1474]: Clock change detected. Flushing caches. Mar 2 14:26:03.127087 systemd-timesyncd[1499]: Contacted time server 10.0.0.1:123 (10.0.0.1). Mar 2 14:26:03.127129 systemd-timesyncd[1499]: Initial clock synchronization to Mon 2026-03-02 14:26:03.127007 UTC. Mar 2 14:26:03.159661 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 2 14:26:03.169751 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Mar 2 14:26:03.180835 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 14:26:03.191287 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 2 14:26:03.218509 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 2 14:26:03.224140 systemd[1]: Reached target network.target - Network. Mar 2 14:26:03.231048 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 14:26:03.237919 systemd[1]: Reached target basic.target - Basic System. Mar 2 14:26:03.249346 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 2 14:26:03.249425 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 2 14:26:03.254815 systemd[1]: Starting containerd.service - containerd container runtime... Mar 2 14:26:03.266946 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 2 14:26:03.283300 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 2 14:26:03.294008 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 2 14:26:03.317497 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 2 14:26:03.321712 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 2 14:26:03.326848 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Mar 2 14:26:03.331073 jq[1538]: false Mar 2 14:26:03.338786 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 2 14:26:03.358610 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 2 14:26:03.373705 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 2 14:26:03.377953 oslogin_cache_refresh[1540]: Refreshing passwd entry cache Mar 2 14:26:03.379020 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Refreshing passwd entry cache Mar 2 14:26:03.389954 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 2 14:26:03.393595 extend-filesystems[1539]: Found /dev/vda6 Mar 2 14:26:03.416179 extend-filesystems[1539]: Found /dev/vda9 Mar 2 14:26:03.412841 oslogin_cache_refresh[1540]: Failure getting users, quitting Mar 2 14:26:03.412622 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 2 14:26:03.425732 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Failure getting users, quitting Mar 2 14:26:03.425732 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Mar 2 14:26:03.425732 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Refreshing group entry cache Mar 2 14:26:03.425808 extend-filesystems[1539]: Checking size of /dev/vda9 Mar 2 14:26:03.412904 oslogin_cache_refresh[1540]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Mar 2 14:26:03.412976 oslogin_cache_refresh[1540]: Refreshing group entry cache Mar 2 14:26:03.438583 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Failure getting groups, quitting Mar 2 14:26:03.438583 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Mar 2 14:26:03.436939 oslogin_cache_refresh[1540]: Failure getting groups, quitting Mar 2 14:26:03.436959 oslogin_cache_refresh[1540]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Mar 2 14:26:03.455729 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Mar 2 14:26:03.476512 extend-filesystems[1539]: Resized partition /dev/vda9 Mar 2 14:26:03.477178 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 2 14:26:03.492109 extend-filesystems[1564]: resize2fs 1.47.3 (8-Jul-2025) Mar 2 14:26:03.518331 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Mar 2 14:26:03.501246 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 2 14:26:03.502318 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 2 14:26:03.507072 systemd[1]: Starting update-engine.service - Update Engine... Mar 2 14:26:03.537742 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 2 14:26:03.566079 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 2 14:26:03.586067 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 2 14:26:03.586422 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 2 14:26:03.587784 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Mar 2 14:26:03.588152 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Mar 2 14:26:03.597958 systemd[1]: motdgen.service: Deactivated successfully. Mar 2 14:26:03.646783 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Mar 2 14:26:03.646836 update_engine[1566]: I20260302 14:26:03.632318 1566 main.cc:92] Flatcar Update Engine starting Mar 2 14:26:03.598313 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 2 14:26:03.647229 jq[1567]: true Mar 2 14:26:03.618838 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 2 14:26:03.619211 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 2 14:26:03.653754 extend-filesystems[1564]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 2 14:26:03.653754 extend-filesystems[1564]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 2 14:26:03.653754 extend-filesystems[1564]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Mar 2 14:26:03.697808 extend-filesystems[1539]: Resized filesystem in /dev/vda9 Mar 2 14:26:03.656992 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 2 14:26:03.666973 (ntainerd)[1574]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 2 14:26:03.670729 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 2 14:26:03.692239 systemd-logind[1554]: Watching system buttons on /dev/input/event2 (Power Button) Mar 2 14:26:03.716796 jq[1575]: true Mar 2 14:26:03.692268 systemd-logind[1554]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 2 14:26:03.693730 systemd-logind[1554]: New seat seat0. Mar 2 14:26:03.700145 systemd[1]: Started systemd-logind.service - User Login Management. Mar 2 14:26:03.757826 tar[1573]: linux-amd64/LICENSE Mar 2 14:26:03.767803 tar[1573]: linux-amd64/helm Mar 2 14:26:03.779014 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Mar 2 14:26:03.811074 dbus-daemon[1536]: [system] SELinux support is enabled Mar 2 14:26:03.813815 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 2 14:26:03.822037 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 2 14:26:03.822102 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 2 14:26:03.833125 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 2 14:26:03.833193 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 2 14:26:03.844778 dbus-daemon[1536]: [system] Successfully activated service 'org.freedesktop.systemd1' Mar 2 14:26:03.847021 update_engine[1566]: I20260302 14:26:03.846962 1566 update_check_scheduler.cc:74] Next update check in 7m26s Mar 2 14:26:03.848547 systemd[1]: Started update-engine.service - Update Engine. Mar 2 14:26:03.864962 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 2 14:26:03.899695 bash[1605]: Updated "/home/core/.ssh/authorized_keys" Mar 2 14:26:03.902419 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 2 14:26:03.910186 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 2 14:26:03.912955 kernel: EDAC MC: Ver: 3.0.0 Mar 2 14:26:03.972647 locksmithd[1606]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 2 14:26:04.073927 containerd[1574]: time="2026-03-02T14:26:04Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Mar 2 14:26:04.076248 containerd[1574]: time="2026-03-02T14:26:04.076159206Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Mar 2 14:26:04.097399 containerd[1574]: time="2026-03-02T14:26:04.097225144Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="11.171µs" Mar 2 14:26:04.097399 containerd[1574]: time="2026-03-02T14:26:04.097308771Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Mar 2 14:26:04.097399 containerd[1574]: time="2026-03-02T14:26:04.097332565Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Mar 2 14:26:04.097730 containerd[1574]: time="2026-03-02T14:26:04.097665226Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Mar 2 14:26:04.097730 containerd[1574]: time="2026-03-02T14:26:04.097690293Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Mar 2 14:26:04.097730 containerd[1574]: time="2026-03-02T14:26:04.097722864Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 2 14:26:04.097850 containerd[1574]: time="2026-03-02T14:26:04.097801541Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 2 14:26:04.097850 containerd[1574]: time="2026-03-02T14:26:04.097816769Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 2 14:26:04.098262 containerd[1574]: time="2026-03-02T14:26:04.098150321Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 2 14:26:04.098262 containerd[1574]: time="2026-03-02T14:26:04.098203792Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 2 14:26:04.098262 containerd[1574]: time="2026-03-02T14:26:04.098219160Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 2 14:26:04.098262 containerd[1574]: time="2026-03-02T14:26:04.098230020Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Mar 2 14:26:04.098390 containerd[1574]: time="2026-03-02T14:26:04.098355285Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Mar 2 14:26:04.098861 containerd[1574]: time="2026-03-02T14:26:04.098745573Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 2 14:26:04.098861 containerd[1574]: time="2026-03-02T14:26:04.098816185Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 2 14:26:04.098861 containerd[1574]: time="2026-03-02T14:26:04.098830261Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Mar 2 14:26:04.098979 containerd[1574]: time="2026-03-02T14:26:04.098876878Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Mar 2 14:26:04.099209 containerd[1574]: time="2026-03-02T14:26:04.099107629Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Mar 2 14:26:04.099255 containerd[1574]: time="2026-03-02T14:26:04.099218095Z" level=info msg="metadata content store policy set" policy=shared Mar 2 14:26:04.109880 systemd-networkd[1473]: eth0: Gained IPv6LL Mar 2 14:26:04.115968 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.118904970Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.118979388Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.119002782Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.119014985Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.119028700Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.119037486Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.119050481Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.119060890Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.119070147Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.120516697Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.120543707Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.120598209Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.120856270Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Mar 2 14:26:04.124746 containerd[1574]: time="2026-03-02T14:26:04.120886547Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Mar 2 14:26:04.124725 systemd[1]: Reached target network-online.target - Network is Online. Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.120915772Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.120934286Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.120950717Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.120966086Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.120979671Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.120999318Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.121014005Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.121026920Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.121039132Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.121112259Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.121141093Z" level=info msg="Start snapshots syncer" Mar 2 14:26:04.134859 containerd[1574]: time="2026-03-02T14:26:04.121174254Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Mar 2 14:26:04.135207 containerd[1574]: time="2026-03-02T14:26:04.125900147Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Mar 2 14:26:04.135207 containerd[1574]: time="2026-03-02T14:26:04.126257284Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.126517269Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131381350Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131411086Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131526431Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131641306Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131660642Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131677874Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131773713Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131886543Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131906581Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.131921980Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.132144726Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.132166897Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 2 14:26:04.135406 containerd[1574]: time="2026-03-02T14:26:04.134367926Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134773764Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134790565Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134807076Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134835620Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134857640Z" level=info msg="runtime interface created" Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134866166Z" level=info msg="created NRI interface" Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134880303Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134900982Z" level=info msg="Connect containerd service" Mar 2 14:26:04.135923 containerd[1574]: time="2026-03-02T14:26:04.134928683Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 2 14:26:04.144816 containerd[1574]: time="2026-03-02T14:26:04.138984495Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 2 14:26:04.136787 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Mar 2 14:26:04.155832 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:26:04.197991 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 2 14:26:04.285301 systemd[1]: coreos-metadata.service: Deactivated successfully. Mar 2 14:26:04.285769 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Mar 2 14:26:04.293024 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 2 14:26:04.325427 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 2 14:26:04.427001 tar[1573]: linux-amd64/README.md Mar 2 14:26:04.425891 systemd[1]: Started containerd.service - containerd container runtime. Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.420529707Z" level=info msg="Start subscribing containerd event" Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.424722214Z" level=info msg="Start recovering state" Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.424857137Z" level=info msg="Start event monitor" Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.424881813Z" level=info msg="Start cni network conf syncer for default" Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.424898574Z" level=info msg="Start streaming server" Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.424918371Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.424928199Z" level=info msg="runtime interface starting up..." Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.424943097Z" level=info msg="starting plugins..." Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.424962363Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.425654285Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.425726390Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 2 14:26:04.427167 containerd[1574]: time="2026-03-02T14:26:04.425828760Z" level=info msg="containerd successfully booted in 0.352865s" Mar 2 14:26:04.465746 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 2 14:26:04.883750 sshd_keygen[1565]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 2 14:26:04.928927 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 2 14:26:04.936326 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 2 14:26:04.944536 systemd[1]: Started sshd@0-10.0.0.9:22-10.0.0.1:58388.service - OpenSSH per-connection server daemon (10.0.0.1:58388). Mar 2 14:26:04.974214 systemd[1]: issuegen.service: Deactivated successfully. Mar 2 14:26:04.974663 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 2 14:26:04.983928 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 2 14:26:05.020850 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 2 14:26:05.030647 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 2 14:26:05.037894 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 2 14:26:05.042839 systemd[1]: Reached target getty.target - Login Prompts. Mar 2 14:26:05.118207 sshd[1666]: Accepted publickey for core from 10.0.0.1 port 58388 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:05.120780 sshd-session[1666]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:05.141753 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 2 14:26:05.154092 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 2 14:26:05.189300 systemd-logind[1554]: New session 1 of user core. Mar 2 14:26:05.213011 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 2 14:26:05.229075 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 2 14:26:05.258745 (systemd)[1678]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 2 14:26:05.266666 systemd-logind[1554]: New session c1 of user core. Mar 2 14:26:05.542107 systemd[1678]: Queued start job for default target default.target. Mar 2 14:26:05.551518 systemd[1678]: Created slice app.slice - User Application Slice. Mar 2 14:26:05.551607 systemd[1678]: Reached target paths.target - Paths. Mar 2 14:26:05.551683 systemd[1678]: Reached target timers.target - Timers. Mar 2 14:26:05.555778 systemd[1678]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 2 14:26:05.589964 systemd[1678]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 2 14:26:05.591303 systemd[1678]: Reached target sockets.target - Sockets. Mar 2 14:26:05.591495 systemd[1678]: Reached target basic.target - Basic System. Mar 2 14:26:05.591616 systemd[1678]: Reached target default.target - Main User Target. Mar 2 14:26:05.591673 systemd[1678]: Startup finished in 298ms. Mar 2 14:26:05.591693 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 2 14:26:05.606741 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 2 14:26:05.660378 systemd[1]: Started sshd@1-10.0.0.9:22-10.0.0.1:58402.service - OpenSSH per-connection server daemon (10.0.0.1:58402). Mar 2 14:26:05.783071 sshd[1689]: Accepted publickey for core from 10.0.0.1 port 58402 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:05.785983 sshd-session[1689]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:05.798417 systemd-logind[1554]: New session 2 of user core. Mar 2 14:26:05.820912 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 2 14:26:05.870657 sshd[1696]: Connection closed by 10.0.0.1 port 58402 Mar 2 14:26:05.864045 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:26:05.862058 sshd-session[1689]: pam_unix(sshd:session): session closed for user core Mar 2 14:26:05.881410 (kubelet)[1698]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 14:26:05.887411 systemd[1]: sshd@1-10.0.0.9:22-10.0.0.1:58402.service: Deactivated successfully. Mar 2 14:26:05.891891 systemd[1]: session-2.scope: Deactivated successfully. Mar 2 14:26:05.895552 systemd-logind[1554]: Session 2 logged out. Waiting for processes to exit. Mar 2 14:26:05.897884 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 2 14:26:05.910121 systemd[1]: Started sshd@2-10.0.0.9:22-10.0.0.1:58408.service - OpenSSH per-connection server daemon (10.0.0.1:58408). Mar 2 14:26:05.929168 systemd[1]: Startup finished in 7.621s (kernel) + 15.052s (initrd) + 10.454s (userspace) = 33.128s. Mar 2 14:26:05.930800 systemd-logind[1554]: Removed session 2. Mar 2 14:26:06.011349 sshd[1704]: Accepted publickey for core from 10.0.0.1 port 58408 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:06.018912 sshd-session[1704]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:06.030036 systemd-logind[1554]: New session 3 of user core. Mar 2 14:26:06.039035 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 2 14:26:06.076803 sshd[1716]: Connection closed by 10.0.0.1 port 58408 Mar 2 14:26:06.078542 sshd-session[1704]: pam_unix(sshd:session): session closed for user core Mar 2 14:26:06.097726 systemd[1]: sshd@2-10.0.0.9:22-10.0.0.1:58408.service: Deactivated successfully. Mar 2 14:26:06.100128 systemd[1]: session-3.scope: Deactivated successfully. Mar 2 14:26:06.102498 systemd-logind[1554]: Session 3 logged out. Waiting for processes to exit. Mar 2 14:26:06.106757 systemd-logind[1554]: Removed session 3. Mar 2 14:26:07.328327 kubelet[1698]: E0302 14:26:07.327741 1698 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 14:26:07.337257 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 14:26:07.339030 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 14:26:07.340844 systemd[1]: kubelet.service: Consumed 1.341s CPU time, 258.7M memory peak. Mar 2 14:26:16.110771 systemd[1]: Started sshd@3-10.0.0.9:22-10.0.0.1:33328.service - OpenSSH per-connection server daemon (10.0.0.1:33328). Mar 2 14:26:16.238341 sshd[1725]: Accepted publickey for core from 10.0.0.1 port 33328 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:16.241381 sshd-session[1725]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:16.275112 systemd-logind[1554]: New session 4 of user core. Mar 2 14:26:16.291055 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 2 14:26:16.336853 sshd[1728]: Connection closed by 10.0.0.1 port 33328 Mar 2 14:26:16.337969 sshd-session[1725]: pam_unix(sshd:session): session closed for user core Mar 2 14:26:16.355015 systemd[1]: sshd@3-10.0.0.9:22-10.0.0.1:33328.service: Deactivated successfully. Mar 2 14:26:16.359328 systemd[1]: session-4.scope: Deactivated successfully. Mar 2 14:26:16.369980 systemd-logind[1554]: Session 4 logged out. Waiting for processes to exit. Mar 2 14:26:16.384626 systemd[1]: Started sshd@4-10.0.0.9:22-10.0.0.1:33338.service - OpenSSH per-connection server daemon (10.0.0.1:33338). Mar 2 14:26:16.390503 systemd-logind[1554]: Removed session 4. Mar 2 14:26:16.488274 sshd[1734]: Accepted publickey for core from 10.0.0.1 port 33338 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:16.491098 sshd-session[1734]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:16.509559 systemd-logind[1554]: New session 5 of user core. Mar 2 14:26:16.517765 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 2 14:26:16.549886 sshd[1737]: Connection closed by 10.0.0.1 port 33338 Mar 2 14:26:16.550868 sshd-session[1734]: pam_unix(sshd:session): session closed for user core Mar 2 14:26:16.563956 systemd[1]: sshd@4-10.0.0.9:22-10.0.0.1:33338.service: Deactivated successfully. Mar 2 14:26:16.567533 systemd[1]: session-5.scope: Deactivated successfully. Mar 2 14:26:16.574501 systemd-logind[1554]: Session 5 logged out. Waiting for processes to exit. Mar 2 14:26:16.582031 systemd[1]: Started sshd@5-10.0.0.9:22-10.0.0.1:33344.service - OpenSSH per-connection server daemon (10.0.0.1:33344). Mar 2 14:26:16.588793 systemd-logind[1554]: Removed session 5. Mar 2 14:26:16.679421 sshd[1743]: Accepted publickey for core from 10.0.0.1 port 33344 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:16.680007 sshd-session[1743]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:16.697020 systemd-logind[1554]: New session 6 of user core. Mar 2 14:26:16.711683 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 2 14:26:16.745495 sshd[1746]: Connection closed by 10.0.0.1 port 33344 Mar 2 14:26:16.746431 sshd-session[1743]: pam_unix(sshd:session): session closed for user core Mar 2 14:26:16.768400 systemd[1]: sshd@5-10.0.0.9:22-10.0.0.1:33344.service: Deactivated successfully. Mar 2 14:26:16.772794 systemd[1]: session-6.scope: Deactivated successfully. Mar 2 14:26:16.778776 systemd-logind[1554]: Session 6 logged out. Waiting for processes to exit. Mar 2 14:26:16.783823 systemd[1]: Started sshd@6-10.0.0.9:22-10.0.0.1:33346.service - OpenSSH per-connection server daemon (10.0.0.1:33346). Mar 2 14:26:16.799028 systemd-logind[1554]: Removed session 6. Mar 2 14:26:16.921900 sshd[1752]: Accepted publickey for core from 10.0.0.1 port 33346 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:16.924300 sshd-session[1752]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:16.946166 systemd-logind[1554]: New session 7 of user core. Mar 2 14:26:16.960765 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 2 14:26:17.015254 sudo[1756]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 2 14:26:17.016764 sudo[1756]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 14:26:17.056709 sudo[1756]: pam_unix(sudo:session): session closed for user root Mar 2 14:26:17.060672 sshd[1755]: Connection closed by 10.0.0.1 port 33346 Mar 2 14:26:17.063313 sshd-session[1752]: pam_unix(sshd:session): session closed for user core Mar 2 14:26:17.080256 systemd[1]: sshd@6-10.0.0.9:22-10.0.0.1:33346.service: Deactivated successfully. Mar 2 14:26:17.084917 systemd[1]: session-7.scope: Deactivated successfully. Mar 2 14:26:17.093939 systemd-logind[1554]: Session 7 logged out. Waiting for processes to exit. Mar 2 14:26:17.096318 systemd[1]: Started sshd@7-10.0.0.9:22-10.0.0.1:33358.service - OpenSSH per-connection server daemon (10.0.0.1:33358). Mar 2 14:26:17.100159 systemd-logind[1554]: Removed session 7. Mar 2 14:26:17.207778 sshd[1762]: Accepted publickey for core from 10.0.0.1 port 33358 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:17.209731 sshd-session[1762]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:17.227265 systemd-logind[1554]: New session 8 of user core. Mar 2 14:26:17.246116 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 2 14:26:17.275383 sudo[1767]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 2 14:26:17.275944 sudo[1767]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 14:26:17.298724 sudo[1767]: pam_unix(sudo:session): session closed for user root Mar 2 14:26:17.311558 sudo[1766]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 2 14:26:17.313783 sudo[1766]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 14:26:17.338323 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 2 14:26:17.345905 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 2 14:26:17.357963 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:26:17.459344 augenrules[1792]: No rules Mar 2 14:26:17.460513 systemd[1]: audit-rules.service: Deactivated successfully. Mar 2 14:26:17.460936 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 2 14:26:17.463725 sudo[1766]: pam_unix(sudo:session): session closed for user root Mar 2 14:26:17.469138 sshd[1765]: Connection closed by 10.0.0.1 port 33358 Mar 2 14:26:17.467803 sshd-session[1762]: pam_unix(sshd:session): session closed for user core Mar 2 14:26:17.488038 systemd[1]: sshd@7-10.0.0.9:22-10.0.0.1:33358.service: Deactivated successfully. Mar 2 14:26:17.498226 systemd[1]: session-8.scope: Deactivated successfully. Mar 2 14:26:17.501846 systemd-logind[1554]: Session 8 logged out. Waiting for processes to exit. Mar 2 14:26:17.511349 systemd[1]: Started sshd@8-10.0.0.9:22-10.0.0.1:33364.service - OpenSSH per-connection server daemon (10.0.0.1:33364). Mar 2 14:26:17.517273 systemd-logind[1554]: Removed session 8. Mar 2 14:26:17.643190 sshd[1801]: Accepted publickey for core from 10.0.0.1 port 33364 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:26:17.644399 sshd-session[1801]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:26:17.657866 systemd-logind[1554]: New session 9 of user core. Mar 2 14:26:17.671893 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 2 14:26:17.706823 sudo[1807]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 2 14:26:17.707273 sudo[1807]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 14:26:17.751933 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:26:17.775634 (kubelet)[1815]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 14:26:18.505082 kubelet[1815]: E0302 14:26:18.503763 1815 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 14:26:18.718833 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 14:26:18.719227 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 14:26:18.720571 systemd[1]: kubelet.service: Consumed 767ms CPU time, 110.8M memory peak. Mar 2 14:26:20.469169 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 2 14:26:20.503362 (dockerd)[1839]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 2 14:26:21.086495 dockerd[1839]: time="2026-03-02T14:26:21.086218077Z" level=info msg="Starting up" Mar 2 14:26:21.093491 dockerd[1839]: time="2026-03-02T14:26:21.092398125Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Mar 2 14:26:21.145565 dockerd[1839]: time="2026-03-02T14:26:21.143094963Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Mar 2 14:26:21.324252 systemd[1]: var-lib-docker-metacopy\x2dcheck1023591823-merged.mount: Deactivated successfully. Mar 2 14:26:21.376086 dockerd[1839]: time="2026-03-02T14:26:21.375517403Z" level=info msg="Loading containers: start." Mar 2 14:26:21.409108 kernel: Initializing XFRM netlink socket Mar 2 14:26:22.490031 systemd-networkd[1473]: docker0: Link UP Mar 2 14:26:22.508666 dockerd[1839]: time="2026-03-02T14:26:22.508403496Z" level=info msg="Loading containers: done." Mar 2 14:26:22.566345 dockerd[1839]: time="2026-03-02T14:26:22.566246444Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 2 14:26:22.566632 dockerd[1839]: time="2026-03-02T14:26:22.566379973Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Mar 2 14:26:22.566632 dockerd[1839]: time="2026-03-02T14:26:22.566560821Z" level=info msg="Initializing buildkit" Mar 2 14:26:23.020760 dockerd[1839]: time="2026-03-02T14:26:23.020268838Z" level=info msg="Completed buildkit initialization" Mar 2 14:26:23.079213 dockerd[1839]: time="2026-03-02T14:26:23.076528499Z" level=info msg="Daemon has completed initialization" Mar 2 14:26:23.079213 dockerd[1839]: time="2026-03-02T14:26:23.078738406Z" level=info msg="API listen on /run/docker.sock" Mar 2 14:26:23.080181 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 2 14:26:26.362699 containerd[1574]: time="2026-03-02T14:26:26.362211057Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.5\"" Mar 2 14:26:27.961113 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount108518412.mount: Deactivated successfully. Mar 2 14:26:29.009866 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 2 14:26:29.037545 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:26:30.138237 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:26:30.201399 (kubelet)[2096]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 14:26:30.803121 kubelet[2096]: E0302 14:26:30.802912 2096 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 14:26:30.810302 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 14:26:30.810667 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 14:26:30.811381 systemd[1]: kubelet.service: Consumed 1.115s CPU time, 110.7M memory peak. Mar 2 14:26:33.631956 containerd[1574]: time="2026-03-02T14:26:33.628552874Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:33.636086 containerd[1574]: time="2026-03-02T14:26:33.634347693Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.34.5: active requests=0, bytes read=27074497" Mar 2 14:26:33.637979 containerd[1574]: time="2026-03-02T14:26:33.637894267Z" level=info msg="ImageCreate event name:\"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:33.652888 containerd[1574]: time="2026-03-02T14:26:33.652180177Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c548633fcd3b4aad59b70815be4c8be54a0fddaddc3fcffa9371eedb0e96417a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:33.656424 containerd[1574]: time="2026-03-02T14:26:33.655973259Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.34.5\" with image id \"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\", repo tag \"registry.k8s.io/kube-apiserver:v1.34.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c548633fcd3b4aad59b70815be4c8be54a0fddaddc3fcffa9371eedb0e96417a\", size \"27071096\" in 7.293600039s" Mar 2 14:26:33.656424 containerd[1574]: time="2026-03-02T14:26:33.656025006Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.5\" returns image reference \"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\"" Mar 2 14:26:33.660977 containerd[1574]: time="2026-03-02T14:26:33.660163512Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.5\"" Mar 2 14:26:36.623240 containerd[1574]: time="2026-03-02T14:26:36.623019412Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:36.627837 containerd[1574]: time="2026-03-02T14:26:36.626618882Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.34.5: active requests=0, bytes read=21165823" Mar 2 14:26:36.633864 containerd[1574]: time="2026-03-02T14:26:36.631494744Z" level=info msg="ImageCreate event name:\"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:36.643090 containerd[1574]: time="2026-03-02T14:26:36.638194682Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f0426100c873816560c520d542fa28999a98dad909edd04365f3b0eead790da3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:36.643090 containerd[1574]: time="2026-03-02T14:26:36.642806772Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.34.5\" with image id \"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\", repo tag \"registry.k8s.io/kube-controller-manager:v1.34.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f0426100c873816560c520d542fa28999a98dad909edd04365f3b0eead790da3\", size \"22822771\" in 2.982541359s" Mar 2 14:26:36.643090 containerd[1574]: time="2026-03-02T14:26:36.642854731Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.5\" returns image reference \"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\"" Mar 2 14:26:36.647945 containerd[1574]: time="2026-03-02T14:26:36.647641638Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.5\"" Mar 2 14:26:38.961061 containerd[1574]: time="2026-03-02T14:26:38.960865144Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:38.964303 containerd[1574]: time="2026-03-02T14:26:38.963923282Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.34.5: active requests=0, bytes read=15729824" Mar 2 14:26:38.967938 containerd[1574]: time="2026-03-02T14:26:38.967774965Z" level=info msg="ImageCreate event name:\"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:38.973526 containerd[1574]: time="2026-03-02T14:26:38.973410995Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:b67b0d627c8e99ffa362bd4d9a60ca9a6c449e363a5f88d2aa8c224bd84ca51d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:38.976285 containerd[1574]: time="2026-03-02T14:26:38.974988731Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.34.5\" with image id \"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\", repo tag \"registry.k8s.io/kube-scheduler:v1.34.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:b67b0d627c8e99ffa362bd4d9a60ca9a6c449e363a5f88d2aa8c224bd84ca51d\", size \"17386790\" in 2.327301037s" Mar 2 14:26:38.976285 containerd[1574]: time="2026-03-02T14:26:38.975060375Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.5\" returns image reference \"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\"" Mar 2 14:26:38.979343 containerd[1574]: time="2026-03-02T14:26:38.978819211Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.5\"" Mar 2 14:26:41.020749 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1097925675.mount: Deactivated successfully. Mar 2 14:26:41.026730 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 2 14:26:41.032787 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:26:41.525746 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:26:41.549077 (kubelet)[2160]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 14:26:41.735396 kubelet[2160]: E0302 14:26:41.732934 2160 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 14:26:41.742700 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 14:26:41.743247 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 14:26:41.745369 systemd[1]: kubelet.service: Consumed 327ms CPU time, 110.3M memory peak. Mar 2 14:26:42.905394 containerd[1574]: time="2026-03-02T14:26:42.904372060Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:42.909855 containerd[1574]: time="2026-03-02T14:26:42.909213790Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.34.5: active requests=0, bytes read=25861770" Mar 2 14:26:42.912362 containerd[1574]: time="2026-03-02T14:26:42.912306102Z" level=info msg="ImageCreate event name:\"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:42.917930 containerd[1574]: time="2026-03-02T14:26:42.917705137Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8a22a3bf452d07af3b5a3064b089d2ad6579d5dd3b850386e05cc0f36dc3f4cf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:42.920753 containerd[1574]: time="2026-03-02T14:26:42.920258790Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.34.5\" with image id \"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\", repo tag \"registry.k8s.io/kube-proxy:v1.34.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:8a22a3bf452d07af3b5a3064b089d2ad6579d5dd3b850386e05cc0f36dc3f4cf\", size \"25860789\" in 3.941404322s" Mar 2 14:26:42.920753 containerd[1574]: time="2026-03-02T14:26:42.920304004Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.5\" returns image reference \"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\"" Mar 2 14:26:42.922514 containerd[1574]: time="2026-03-02T14:26:42.922048310Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\"" Mar 2 14:26:43.542213 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3216208860.mount: Deactivated successfully. Mar 2 14:26:46.604551 containerd[1574]: time="2026-03-02T14:26:46.604183847Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:46.609645 containerd[1574]: time="2026-03-02T14:26:46.609348727Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.1: active requests=0, bytes read=22388007" Mar 2 14:26:46.617199 containerd[1574]: time="2026-03-02T14:26:46.617026260Z" level=info msg="ImageCreate event name:\"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:46.624240 containerd[1574]: time="2026-03-02T14:26:46.623998106Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:46.627759 containerd[1574]: time="2026-03-02T14:26:46.626615803Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.1\" with image id \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\", size \"22384805\" in 3.704510547s" Mar 2 14:26:46.627759 containerd[1574]: time="2026-03-02T14:26:46.626653704Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\" returns image reference \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\"" Mar 2 14:26:46.627759 containerd[1574]: time="2026-03-02T14:26:46.627544643Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Mar 2 14:26:47.203768 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3919591560.mount: Deactivated successfully. Mar 2 14:26:47.224993 containerd[1574]: time="2026-03-02T14:26:47.222886177Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:47.228613 containerd[1574]: time="2026-03-02T14:26:47.228087337Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=321218" Mar 2 14:26:47.239518 containerd[1574]: time="2026-03-02T14:26:47.237894109Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:47.246951 containerd[1574]: time="2026-03-02T14:26:47.242802008Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:47.246951 containerd[1574]: time="2026-03-02T14:26:47.244615867Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 617.040367ms" Mar 2 14:26:47.246951 containerd[1574]: time="2026-03-02T14:26:47.244647576Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Mar 2 14:26:47.250499 containerd[1574]: time="2026-03-02T14:26:47.249034243Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\"" Mar 2 14:26:47.916740 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1651164785.mount: Deactivated successfully. Mar 2 14:26:48.801791 update_engine[1566]: I20260302 14:26:48.800647 1566 update_attempter.cc:509] Updating boot flags... Mar 2 14:26:50.692375 containerd[1574]: time="2026-03-02T14:26:50.692146048Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.5-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:50.695272 containerd[1574]: time="2026-03-02T14:26:50.695113888Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.5-0: active requests=0, bytes read=22860674" Mar 2 14:26:50.697842 containerd[1574]: time="2026-03-02T14:26:50.697725514Z" level=info msg="ImageCreate event name:\"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:50.705461 containerd[1574]: time="2026-03-02T14:26:50.705338615Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:26:50.707104 containerd[1574]: time="2026-03-02T14:26:50.707048078Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.5-0\" with image id \"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\", repo tag \"registry.k8s.io/etcd:3.6.5-0\", repo digest \"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\", size \"22871747\" in 3.45630814s" Mar 2 14:26:50.707104 containerd[1574]: time="2026-03-02T14:26:50.707080719Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\" returns image reference \"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\"" Mar 2 14:26:51.771313 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 2 14:26:51.773426 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:26:52.149552 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:26:52.172153 (kubelet)[2334]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 14:26:52.296933 kubelet[2334]: E0302 14:26:52.296702 2334 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 14:26:52.303180 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 14:26:52.303414 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 14:26:52.304102 systemd[1]: kubelet.service: Consumed 288ms CPU time, 110.3M memory peak. Mar 2 14:26:57.280196 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:26:57.280544 systemd[1]: kubelet.service: Consumed 288ms CPU time, 110.3M memory peak. Mar 2 14:26:57.287234 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:26:57.380956 systemd[1]: Reload requested from client PID 2350 ('systemctl') (unit session-9.scope)... Mar 2 14:26:57.380998 systemd[1]: Reloading... Mar 2 14:26:57.543506 zram_generator::config[2393]: No configuration found. Mar 2 14:26:58.084178 systemd[1]: Reloading finished in 702 ms. Mar 2 14:26:58.240090 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 2 14:26:58.240293 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 2 14:26:58.240880 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:26:58.242573 systemd[1]: kubelet.service: Consumed 188ms CPU time, 98.1M memory peak. Mar 2 14:26:58.245709 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:26:58.665561 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:26:58.686982 (kubelet)[2440]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 14:26:58.830762 kubelet[2440]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 2 14:26:58.830762 kubelet[2440]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 14:26:58.831281 kubelet[2440]: I0302 14:26:58.830779 2440 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 2 14:26:59.810477 kubelet[2440]: I0302 14:26:59.810280 2440 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Mar 2 14:26:59.810477 kubelet[2440]: I0302 14:26:59.810327 2440 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 14:26:59.810477 kubelet[2440]: I0302 14:26:59.810361 2440 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 2 14:26:59.810477 kubelet[2440]: I0302 14:26:59.810373 2440 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 14:26:59.811218 kubelet[2440]: I0302 14:26:59.810704 2440 server.go:956] "Client rotation is on, will bootstrap in background" Mar 2 14:26:59.920521 kubelet[2440]: I0302 14:26:59.918298 2440 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 14:26:59.920521 kubelet[2440]: E0302 14:26:59.920303 2440 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.9:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 14:26:59.938651 kubelet[2440]: I0302 14:26:59.938511 2440 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 2 14:26:59.957147 kubelet[2440]: I0302 14:26:59.956722 2440 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 2 14:26:59.959064 kubelet[2440]: I0302 14:26:59.958123 2440 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 14:26:59.959064 kubelet[2440]: I0302 14:26:59.958192 2440 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 2 14:26:59.959064 kubelet[2440]: I0302 14:26:59.958409 2440 topology_manager.go:138] "Creating topology manager with none policy" Mar 2 14:26:59.959064 kubelet[2440]: I0302 14:26:59.958421 2440 container_manager_linux.go:306] "Creating device plugin manager" Mar 2 14:26:59.960297 kubelet[2440]: I0302 14:26:59.958632 2440 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Mar 2 14:26:59.966388 kubelet[2440]: I0302 14:26:59.962374 2440 state_mem.go:36] "Initialized new in-memory state store" Mar 2 14:26:59.966388 kubelet[2440]: I0302 14:26:59.962663 2440 kubelet.go:475] "Attempting to sync node with API server" Mar 2 14:26:59.966388 kubelet[2440]: I0302 14:26:59.962682 2440 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 14:26:59.966388 kubelet[2440]: I0302 14:26:59.962711 2440 kubelet.go:387] "Adding apiserver pod source" Mar 2 14:26:59.966388 kubelet[2440]: I0302 14:26:59.962724 2440 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 14:26:59.966840 kubelet[2440]: I0302 14:26:59.966817 2440 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Mar 2 14:26:59.967583 kubelet[2440]: E0302 14:26:59.967523 2440 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.9:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 14:26:59.971017 kubelet[2440]: I0302 14:26:59.969496 2440 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 14:26:59.971017 kubelet[2440]: I0302 14:26:59.969562 2440 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 2 14:26:59.971017 kubelet[2440]: W0302 14:26:59.969639 2440 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 2 14:26:59.971017 kubelet[2440]: E0302 14:26:59.970087 2440 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.9:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 14:26:59.982073 kubelet[2440]: I0302 14:26:59.978743 2440 server.go:1262] "Started kubelet" Mar 2 14:26:59.982073 kubelet[2440]: I0302 14:26:59.980194 2440 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 14:26:59.982073 kubelet[2440]: I0302 14:26:59.980246 2440 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 2 14:26:59.982073 kubelet[2440]: I0302 14:26:59.980552 2440 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 14:26:59.982073 kubelet[2440]: I0302 14:26:59.980677 2440 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 14:26:59.982073 kubelet[2440]: I0302 14:26:59.981793 2440 server.go:310] "Adding debug handlers to kubelet server" Mar 2 14:26:59.985170 kubelet[2440]: I0302 14:26:59.984752 2440 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 2 14:26:59.985170 kubelet[2440]: I0302 14:26:59.985108 2440 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 14:26:59.989075 kubelet[2440]: E0302 14:26:59.987957 2440 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.9:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.9:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18990c73d4d7ced8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 14:26:59.978686168 +0000 UTC m=+1.277809805,LastTimestamp:2026-03-02 14:26:59.978686168 +0000 UTC m=+1.277809805,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 14:26:59.990242 kubelet[2440]: I0302 14:26:59.990221 2440 volume_manager.go:313] "Starting Kubelet Volume Manager" Mar 2 14:26:59.990425 kubelet[2440]: I0302 14:26:59.990407 2440 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 2 14:26:59.990619 kubelet[2440]: I0302 14:26:59.990599 2440 reconciler.go:29] "Reconciler: start to sync state" Mar 2 14:26:59.991201 kubelet[2440]: E0302 14:26:59.991174 2440 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.9:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 14:26:59.991569 kubelet[2440]: E0302 14:26:59.991513 2440 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 14:26:59.992807 kubelet[2440]: E0302 14:26:59.992754 2440 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.9:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.9:6443: connect: connection refused" interval="200ms" Mar 2 14:27:00.003555 kubelet[2440]: I0302 14:27:00.003530 2440 factory.go:223] Registration of the containerd container factory successfully Mar 2 14:27:00.004975 kubelet[2440]: I0302 14:27:00.004954 2440 factory.go:223] Registration of the systemd container factory successfully Mar 2 14:27:00.005208 kubelet[2440]: I0302 14:27:00.005181 2440 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 14:27:00.006505 kubelet[2440]: E0302 14:27:00.006317 2440 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 14:27:00.040531 kubelet[2440]: I0302 14:27:00.040376 2440 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 2 14:27:00.040531 kubelet[2440]: I0302 14:27:00.040430 2440 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 2 14:27:00.040531 kubelet[2440]: I0302 14:27:00.040501 2440 state_mem.go:36] "Initialized new in-memory state store" Mar 2 14:27:00.054191 kubelet[2440]: I0302 14:27:00.053547 2440 policy_none.go:49] "None policy: Start" Mar 2 14:27:00.054191 kubelet[2440]: I0302 14:27:00.053594 2440 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 2 14:27:00.054191 kubelet[2440]: I0302 14:27:00.053615 2440 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 2 14:27:00.065763 kubelet[2440]: I0302 14:27:00.065571 2440 policy_none.go:47] "Start" Mar 2 14:27:00.081816 kubelet[2440]: I0302 14:27:00.081605 2440 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 2 14:27:00.090268 kubelet[2440]: I0302 14:27:00.087083 2440 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 2 14:27:00.090268 kubelet[2440]: I0302 14:27:00.087108 2440 status_manager.go:244] "Starting to sync pod status with apiserver" Mar 2 14:27:00.090268 kubelet[2440]: I0302 14:27:00.087137 2440 kubelet.go:2428] "Starting kubelet main sync loop" Mar 2 14:27:00.090268 kubelet[2440]: E0302 14:27:00.087192 2440 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 14:27:00.090098 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 2 14:27:00.095101 kubelet[2440]: E0302 14:27:00.093718 2440 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 14:27:00.095101 kubelet[2440]: E0302 14:27:00.093969 2440 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.9:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 14:27:00.130553 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 2 14:27:00.143408 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 2 14:27:00.164776 kubelet[2440]: E0302 14:27:00.163612 2440 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 14:27:00.164943 kubelet[2440]: I0302 14:27:00.164802 2440 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 2 14:27:00.164943 kubelet[2440]: I0302 14:27:00.164818 2440 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 14:27:00.165332 kubelet[2440]: I0302 14:27:00.165213 2440 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 2 14:27:00.169690 kubelet[2440]: E0302 14:27:00.169546 2440 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 14:27:00.169690 kubelet[2440]: E0302 14:27:00.169586 2440 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 2 14:27:00.192349 kubelet[2440]: I0302 14:27:00.192308 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a8d35cbde6c3a030da03a196ce305505-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"a8d35cbde6c3a030da03a196ce305505\") " pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:00.192831 kubelet[2440]: I0302 14:27:00.192707 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a8d35cbde6c3a030da03a196ce305505-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"a8d35cbde6c3a030da03a196ce305505\") " pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:00.193233 kubelet[2440]: I0302 14:27:00.192867 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a8d35cbde6c3a030da03a196ce305505-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"a8d35cbde6c3a030da03a196ce305505\") " pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:00.195537 kubelet[2440]: E0302 14:27:00.195363 2440 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.9:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.9:6443: connect: connection refused" interval="400ms" Mar 2 14:27:00.226540 systemd[1]: Created slice kubepods-burstable-poda8d35cbde6c3a030da03a196ce305505.slice - libcontainer container kubepods-burstable-poda8d35cbde6c3a030da03a196ce305505.slice. Mar 2 14:27:00.255816 kubelet[2440]: E0302 14:27:00.254410 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:00.260072 systemd[1]: Created slice kubepods-burstable-poddb0989cdb653dfec284dd4f35625e9e7.slice - libcontainer container kubepods-burstable-poddb0989cdb653dfec284dd4f35625e9e7.slice. Mar 2 14:27:00.267381 kubelet[2440]: E0302 14:27:00.265688 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:00.267781 kubelet[2440]: I0302 14:27:00.267751 2440 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 14:27:00.268272 kubelet[2440]: E0302 14:27:00.268219 2440 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.9:6443/api/v1/nodes\": dial tcp 10.0.0.9:6443: connect: connection refused" node="localhost" Mar 2 14:27:00.271247 systemd[1]: Created slice kubepods-burstable-pod89efda49e166906783d8d868d41ebb86.slice - libcontainer container kubepods-burstable-pod89efda49e166906783d8d868d41ebb86.slice. Mar 2 14:27:00.275351 kubelet[2440]: E0302 14:27:00.274598 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:00.295964 kubelet[2440]: I0302 14:27:00.295851 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:00.295964 kubelet[2440]: I0302 14:27:00.295940 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:00.296178 kubelet[2440]: I0302 14:27:00.295966 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:00.296178 kubelet[2440]: I0302 14:27:00.296021 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/89efda49e166906783d8d868d41ebb86-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"89efda49e166906783d8d868d41ebb86\") " pod="kube-system/kube-scheduler-localhost" Mar 2 14:27:00.296178 kubelet[2440]: I0302 14:27:00.296152 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:00.296284 kubelet[2440]: I0302 14:27:00.296218 2440 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:00.475355 kubelet[2440]: I0302 14:27:00.472791 2440 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 14:27:00.475355 kubelet[2440]: E0302 14:27:00.474417 2440 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.9:6443/api/v1/nodes\": dial tcp 10.0.0.9:6443: connect: connection refused" node="localhost" Mar 2 14:27:00.569683 kubelet[2440]: E0302 14:27:00.568676 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:00.573419 containerd[1574]: time="2026-03-02T14:27:00.573227915Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:a8d35cbde6c3a030da03a196ce305505,Namespace:kube-system,Attempt:0,}" Mar 2 14:27:00.586478 kubelet[2440]: E0302 14:27:00.585007 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:00.589227 containerd[1574]: time="2026-03-02T14:27:00.586799051Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:db0989cdb653dfec284dd4f35625e9e7,Namespace:kube-system,Attempt:0,}" Mar 2 14:27:00.591189 kubelet[2440]: E0302 14:27:00.590130 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:00.591250 containerd[1574]: time="2026-03-02T14:27:00.590554538Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:89efda49e166906783d8d868d41ebb86,Namespace:kube-system,Attempt:0,}" Mar 2 14:27:00.597330 kubelet[2440]: E0302 14:27:00.597045 2440 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.9:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.9:6443: connect: connection refused" interval="800ms" Mar 2 14:27:00.878126 kubelet[2440]: I0302 14:27:00.877699 2440 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 14:27:00.878427 kubelet[2440]: E0302 14:27:00.878397 2440 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.9:6443/api/v1/nodes\": dial tcp 10.0.0.9:6443: connect: connection refused" node="localhost" Mar 2 14:27:00.931683 kubelet[2440]: E0302 14:27:00.930625 2440 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.9:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 14:27:01.121027 kubelet[2440]: E0302 14:27:01.119305 2440 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.9:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 14:27:01.396425 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1924001792.mount: Deactivated successfully. Mar 2 14:27:01.400295 kubelet[2440]: E0302 14:27:01.399832 2440 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.9:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.9:6443: connect: connection refused" interval="1.6s" Mar 2 14:27:01.420766 containerd[1574]: time="2026-03-02T14:27:01.420512117Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 14:27:01.434485 containerd[1574]: time="2026-03-02T14:27:01.434294463Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Mar 2 14:27:01.444431 containerd[1574]: time="2026-03-02T14:27:01.442848462Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 14:27:01.448645 containerd[1574]: time="2026-03-02T14:27:01.448224449Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 14:27:01.452953 containerd[1574]: time="2026-03-02T14:27:01.450309011Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 14:27:01.455079 containerd[1574]: time="2026-03-02T14:27:01.454938273Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Mar 2 14:27:01.458033 containerd[1574]: time="2026-03-02T14:27:01.455937068Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Mar 2 14:27:01.464991 containerd[1574]: time="2026-03-02T14:27:01.460323888Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 14:27:01.464991 containerd[1574]: time="2026-03-02T14:27:01.463111842Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 873.383998ms" Mar 2 14:27:01.474762 containerd[1574]: time="2026-03-02T14:27:01.474722478Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 869.629678ms" Mar 2 14:27:01.487995 containerd[1574]: time="2026-03-02T14:27:01.487375681Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 894.811005ms" Mar 2 14:27:01.493731 kubelet[2440]: E0302 14:27:01.493666 2440 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.9:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 14:27:01.553831 containerd[1574]: time="2026-03-02T14:27:01.553668865Z" level=info msg="connecting to shim b45025ecf2a4016256f3b5752dd43b1dc00848de1bbadbe9f725232e23f21851" address="unix:///run/containerd/s/d3f3d0d1d0955828daeef03ca6dd103e514faf67ebcd4b306b8020ff1668213b" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:27:01.613667 containerd[1574]: time="2026-03-02T14:27:01.612822891Z" level=info msg="connecting to shim 7e1777010bed87983c41956462aa5590efc6a2e54f2b0cbb9c18c3c8381fcc41" address="unix:///run/containerd/s/0244cdc50461984a8a3814afff4d2435d6248e8de9ea0b7d7d0401f02aebde99" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:27:01.617037 containerd[1574]: time="2026-03-02T14:27:01.615842658Z" level=info msg="connecting to shim 962f50155d5c0cac4a96191acea6e5904eb1a472dffa79372b30ceb105a62329" address="unix:///run/containerd/s/ce5bf35529599ff2ec2d5fe8d12d0be9ffeaa58093e4568861ebaaa70fffc582" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:27:01.660286 systemd[1]: Started cri-containerd-b45025ecf2a4016256f3b5752dd43b1dc00848de1bbadbe9f725232e23f21851.scope - libcontainer container b45025ecf2a4016256f3b5752dd43b1dc00848de1bbadbe9f725232e23f21851. Mar 2 14:27:01.680274 kubelet[2440]: I0302 14:27:01.680241 2440 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 14:27:01.680808 kubelet[2440]: E0302 14:27:01.680778 2440 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.9:6443/api/v1/nodes\": dial tcp 10.0.0.9:6443: connect: connection refused" node="localhost" Mar 2 14:27:01.687251 kubelet[2440]: E0302 14:27:01.687166 2440 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.9:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 14:27:01.707653 systemd[1]: Started cri-containerd-962f50155d5c0cac4a96191acea6e5904eb1a472dffa79372b30ceb105a62329.scope - libcontainer container 962f50155d5c0cac4a96191acea6e5904eb1a472dffa79372b30ceb105a62329. Mar 2 14:27:01.748761 systemd[1]: Started cri-containerd-7e1777010bed87983c41956462aa5590efc6a2e54f2b0cbb9c18c3c8381fcc41.scope - libcontainer container 7e1777010bed87983c41956462aa5590efc6a2e54f2b0cbb9c18c3c8381fcc41. Mar 2 14:27:01.972313 containerd[1574]: time="2026-03-02T14:27:01.971367849Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:a8d35cbde6c3a030da03a196ce305505,Namespace:kube-system,Attempt:0,} returns sandbox id \"b45025ecf2a4016256f3b5752dd43b1dc00848de1bbadbe9f725232e23f21851\"" Mar 2 14:27:01.977527 kubelet[2440]: E0302 14:27:01.977415 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:01.991075 containerd[1574]: time="2026-03-02T14:27:01.990815053Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:db0989cdb653dfec284dd4f35625e9e7,Namespace:kube-system,Attempt:0,} returns sandbox id \"962f50155d5c0cac4a96191acea6e5904eb1a472dffa79372b30ceb105a62329\"" Mar 2 14:27:01.991611 kubelet[2440]: E0302 14:27:01.991505 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:02.005072 containerd[1574]: time="2026-03-02T14:27:02.003327371Z" level=info msg="CreateContainer within sandbox \"b45025ecf2a4016256f3b5752dd43b1dc00848de1bbadbe9f725232e23f21851\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 2 14:27:02.019217 containerd[1574]: time="2026-03-02T14:27:02.018677128Z" level=info msg="CreateContainer within sandbox \"962f50155d5c0cac4a96191acea6e5904eb1a472dffa79372b30ceb105a62329\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 2 14:27:02.027950 containerd[1574]: time="2026-03-02T14:27:02.027170926Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:89efda49e166906783d8d868d41ebb86,Namespace:kube-system,Attempt:0,} returns sandbox id \"7e1777010bed87983c41956462aa5590efc6a2e54f2b0cbb9c18c3c8381fcc41\"" Mar 2 14:27:02.031149 kubelet[2440]: E0302 14:27:02.030831 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:02.050323 containerd[1574]: time="2026-03-02T14:27:02.050261614Z" level=info msg="Container b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:27:02.051533 containerd[1574]: time="2026-03-02T14:27:02.051501032Z" level=info msg="CreateContainer within sandbox \"7e1777010bed87983c41956462aa5590efc6a2e54f2b0cbb9c18c3c8381fcc41\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 2 14:27:02.063417 containerd[1574]: time="2026-03-02T14:27:02.062498088Z" level=info msg="Container 833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:27:02.068807 kubelet[2440]: E0302 14:27:02.068566 2440 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.9:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.9:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 14:27:02.085658 containerd[1574]: time="2026-03-02T14:27:02.085611420Z" level=info msg="CreateContainer within sandbox \"b45025ecf2a4016256f3b5752dd43b1dc00848de1bbadbe9f725232e23f21851\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f\"" Mar 2 14:27:02.087765 containerd[1574]: time="2026-03-02T14:27:02.087659283Z" level=info msg="StartContainer for \"b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f\"" Mar 2 14:27:02.092488 containerd[1574]: time="2026-03-02T14:27:02.092229425Z" level=info msg="connecting to shim b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f" address="unix:///run/containerd/s/d3f3d0d1d0955828daeef03ca6dd103e514faf67ebcd4b306b8020ff1668213b" protocol=ttrpc version=3 Mar 2 14:27:02.121980 containerd[1574]: time="2026-03-02T14:27:02.120714394Z" level=info msg="CreateContainer within sandbox \"962f50155d5c0cac4a96191acea6e5904eb1a472dffa79372b30ceb105a62329\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947\"" Mar 2 14:27:02.126495 containerd[1574]: time="2026-03-02T14:27:02.125121984Z" level=info msg="StartContainer for \"833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947\"" Mar 2 14:27:02.131263 containerd[1574]: time="2026-03-02T14:27:02.130023353Z" level=info msg="connecting to shim 833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947" address="unix:///run/containerd/s/ce5bf35529599ff2ec2d5fe8d12d0be9ffeaa58093e4568861ebaaa70fffc582" protocol=ttrpc version=3 Mar 2 14:27:02.144299 containerd[1574]: time="2026-03-02T14:27:02.143135944Z" level=info msg="Container 3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:27:02.169277 systemd[1]: Started cri-containerd-b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f.scope - libcontainer container b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f. Mar 2 14:27:02.173515 containerd[1574]: time="2026-03-02T14:27:02.172223001Z" level=info msg="CreateContainer within sandbox \"7e1777010bed87983c41956462aa5590efc6a2e54f2b0cbb9c18c3c8381fcc41\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc\"" Mar 2 14:27:02.175586 containerd[1574]: time="2026-03-02T14:27:02.175520135Z" level=info msg="StartContainer for \"3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc\"" Mar 2 14:27:02.185331 containerd[1574]: time="2026-03-02T14:27:02.185277431Z" level=info msg="connecting to shim 3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc" address="unix:///run/containerd/s/0244cdc50461984a8a3814afff4d2435d6248e8de9ea0b7d7d0401f02aebde99" protocol=ttrpc version=3 Mar 2 14:27:02.205934 systemd[1]: Started cri-containerd-833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947.scope - libcontainer container 833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947. Mar 2 14:27:02.259173 systemd[1]: Started cri-containerd-3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc.scope - libcontainer container 3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc. Mar 2 14:27:02.408429 containerd[1574]: time="2026-03-02T14:27:02.407640395Z" level=info msg="StartContainer for \"b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f\" returns successfully" Mar 2 14:27:02.494815 containerd[1574]: time="2026-03-02T14:27:02.494720578Z" level=info msg="StartContainer for \"833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947\" returns successfully" Mar 2 14:27:02.556420 containerd[1574]: time="2026-03-02T14:27:02.556323242Z" level=info msg="StartContainer for \"3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc\" returns successfully" Mar 2 14:27:03.145372 kubelet[2440]: E0302 14:27:03.145241 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:03.146684 kubelet[2440]: E0302 14:27:03.145422 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:03.155936 kubelet[2440]: E0302 14:27:03.155822 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:03.156071 kubelet[2440]: E0302 14:27:03.156032 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:03.157250 kubelet[2440]: E0302 14:27:03.157192 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:03.157512 kubelet[2440]: E0302 14:27:03.157388 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:03.291759 kubelet[2440]: I0302 14:27:03.290776 2440 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 14:27:04.166970 kubelet[2440]: E0302 14:27:04.166841 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:04.173336 kubelet[2440]: E0302 14:27:04.167073 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:04.173336 kubelet[2440]: E0302 14:27:04.167383 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:04.173336 kubelet[2440]: E0302 14:27:04.167840 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:04.173995 kubelet[2440]: E0302 14:27:04.173542 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:04.173995 kubelet[2440]: E0302 14:27:04.173751 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:05.177960 kubelet[2440]: E0302 14:27:05.174284 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:05.177960 kubelet[2440]: E0302 14:27:05.174524 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:05.182578 kubelet[2440]: E0302 14:27:05.180987 2440 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 14:27:05.182578 kubelet[2440]: E0302 14:27:05.181165 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:07.176012 kubelet[2440]: E0302 14:27:07.174230 2440 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Mar 2 14:27:07.481094 kubelet[2440]: I0302 14:27:07.477570 2440 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Mar 2 14:27:07.494345 kubelet[2440]: I0302 14:27:07.494138 2440 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 2 14:27:07.534923 kubelet[2440]: E0302 14:27:07.534778 2440 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.18990c73d4d7ced8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 14:26:59.978686168 +0000 UTC m=+1.277809805,LastTimestamp:2026-03-02 14:26:59.978686168 +0000 UTC m=+1.277809805,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 14:27:07.690321 kubelet[2440]: I0302 14:27:07.690128 2440 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:07.811349 kubelet[2440]: E0302 14:27:07.809647 2440 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.18990c73d67d0827 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 14:27:00.006291495 +0000 UTC m=+1.305415153,LastTimestamp:2026-03-02 14:27:00.006291495 +0000 UTC m=+1.305415153,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 14:27:07.891385 kubelet[2440]: E0302 14:27:07.891287 2440 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:07.892354 kubelet[2440]: E0302 14:27:07.892264 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:07.897167 kubelet[2440]: E0302 14:27:07.896516 2440 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Mar 2 14:27:07.897167 kubelet[2440]: I0302 14:27:07.896547 2440 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:07.908592 kubelet[2440]: E0302 14:27:07.908282 2440 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:07.908592 kubelet[2440]: I0302 14:27:07.908316 2440 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:07.910668 kubelet[2440]: E0302 14:27:07.910645 2440 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:07.970229 kubelet[2440]: I0302 14:27:07.969308 2440 apiserver.go:52] "Watching apiserver" Mar 2 14:27:07.992414 kubelet[2440]: I0302 14:27:07.991580 2440 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 2 14:27:08.173561 kubelet[2440]: I0302 14:27:08.173279 2440 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 2 14:27:08.230618 kubelet[2440]: E0302 14:27:08.230516 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:09.189114 kubelet[2440]: E0302 14:27:09.189077 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:10.296058 kubelet[2440]: I0302 14:27:10.293964 2440 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=2.293924801 podStartE2EDuration="2.293924801s" podCreationTimestamp="2026-03-02 14:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 14:27:10.275922812 +0000 UTC m=+11.575046459" watchObservedRunningTime="2026-03-02 14:27:10.293924801 +0000 UTC m=+11.593048438" Mar 2 14:27:11.473992 kubelet[2440]: I0302 14:27:11.470357 2440 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:11.568703 kubelet[2440]: E0302 14:27:11.562489 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:12.220623 kubelet[2440]: E0302 14:27:12.215430 2440 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:15.038160 systemd[1]: Reload requested from client PID 2735 ('systemctl') (unit session-9.scope)... Mar 2 14:27:15.038186 systemd[1]: Reloading... Mar 2 14:27:15.378140 zram_generator::config[2778]: No configuration found. Mar 2 14:27:15.986239 systemd[1]: Reloading finished in 947 ms. Mar 2 14:27:16.070361 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:27:16.106132 systemd[1]: kubelet.service: Deactivated successfully. Mar 2 14:27:16.107087 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:27:16.107201 systemd[1]: kubelet.service: Consumed 2.256s CPU time, 127.3M memory peak. Mar 2 14:27:16.115223 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 14:27:16.520019 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 14:27:16.552735 (kubelet)[2823]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 14:27:16.723154 kubelet[2823]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 2 14:27:16.723154 kubelet[2823]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 14:27:16.723154 kubelet[2823]: I0302 14:27:16.720783 2823 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 2 14:27:16.755355 kubelet[2823]: I0302 14:27:16.755316 2823 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Mar 2 14:27:16.757557 kubelet[2823]: I0302 14:27:16.755653 2823 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 14:27:16.757557 kubelet[2823]: I0302 14:27:16.755701 2823 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 2 14:27:16.757557 kubelet[2823]: I0302 14:27:16.755716 2823 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 14:27:16.757557 kubelet[2823]: I0302 14:27:16.756325 2823 server.go:956] "Client rotation is on, will bootstrap in background" Mar 2 14:27:16.758981 kubelet[2823]: I0302 14:27:16.758963 2823 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 2 14:27:16.779484 kubelet[2823]: I0302 14:27:16.779196 2823 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 14:27:16.804006 kubelet[2823]: I0302 14:27:16.801020 2823 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 2 14:27:16.822005 kubelet[2823]: I0302 14:27:16.820281 2823 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 2 14:27:16.822005 kubelet[2823]: I0302 14:27:16.820788 2823 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 14:27:16.822005 kubelet[2823]: I0302 14:27:16.820830 2823 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 2 14:27:16.822005 kubelet[2823]: I0302 14:27:16.821067 2823 topology_manager.go:138] "Creating topology manager with none policy" Mar 2 14:27:16.822361 kubelet[2823]: I0302 14:27:16.821082 2823 container_manager_linux.go:306] "Creating device plugin manager" Mar 2 14:27:16.822361 kubelet[2823]: I0302 14:27:16.821121 2823 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Mar 2 14:27:16.822361 kubelet[2823]: I0302 14:27:16.821504 2823 state_mem.go:36] "Initialized new in-memory state store" Mar 2 14:27:16.822361 kubelet[2823]: I0302 14:27:16.821851 2823 kubelet.go:475] "Attempting to sync node with API server" Mar 2 14:27:16.824664 kubelet[2823]: I0302 14:27:16.824285 2823 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 14:27:16.824664 kubelet[2823]: I0302 14:27:16.824356 2823 kubelet.go:387] "Adding apiserver pod source" Mar 2 14:27:16.824664 kubelet[2823]: I0302 14:27:16.824371 2823 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 14:27:16.834106 kubelet[2823]: I0302 14:27:16.830684 2823 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Mar 2 14:27:16.851427 kubelet[2823]: I0302 14:27:16.845149 2823 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 14:27:16.851427 kubelet[2823]: I0302 14:27:16.845195 2823 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 2 14:27:16.924317 kubelet[2823]: I0302 14:27:16.923813 2823 server.go:1262] "Started kubelet" Mar 2 14:27:16.932380 kubelet[2823]: I0302 14:27:16.927839 2823 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 14:27:16.932380 kubelet[2823]: I0302 14:27:16.929139 2823 server.go:310] "Adding debug handlers to kubelet server" Mar 2 14:27:16.932380 kubelet[2823]: I0302 14:27:16.929730 2823 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 14:27:16.932380 kubelet[2823]: I0302 14:27:16.929773 2823 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 2 14:27:16.937661 kubelet[2823]: I0302 14:27:16.935725 2823 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 14:27:16.954578 kubelet[2823]: I0302 14:27:16.943633 2823 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 2 14:27:16.954578 kubelet[2823]: I0302 14:27:16.953333 2823 volume_manager.go:313] "Starting Kubelet Volume Manager" Mar 2 14:27:16.954578 kubelet[2823]: I0302 14:27:16.953636 2823 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 14:27:16.961051 kubelet[2823]: I0302 14:27:16.959903 2823 reconciler.go:29] "Reconciler: start to sync state" Mar 2 14:27:16.961051 kubelet[2823]: I0302 14:27:16.959977 2823 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 2 14:27:16.965150 kubelet[2823]: I0302 14:27:16.965038 2823 factory.go:223] Registration of the systemd container factory successfully Mar 2 14:27:16.965381 kubelet[2823]: I0302 14:27:16.965355 2823 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 14:27:16.972499 kubelet[2823]: E0302 14:27:16.966008 2823 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 14:27:16.974304 kubelet[2823]: I0302 14:27:16.974192 2823 factory.go:223] Registration of the containerd container factory successfully Mar 2 14:27:17.023322 kubelet[2823]: I0302 14:27:17.017162 2823 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 2 14:27:17.097818 kubelet[2823]: I0302 14:27:17.097658 2823 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 2 14:27:17.097818 kubelet[2823]: I0302 14:27:17.097688 2823 status_manager.go:244] "Starting to sync pod status with apiserver" Mar 2 14:27:17.097818 kubelet[2823]: I0302 14:27:17.097720 2823 kubelet.go:2428] "Starting kubelet main sync loop" Mar 2 14:27:17.099405 kubelet[2823]: E0302 14:27:17.097777 2823 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 14:27:17.149255 kubelet[2823]: I0302 14:27:17.148289 2823 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 2 14:27:17.149255 kubelet[2823]: I0302 14:27:17.148312 2823 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 2 14:27:17.149255 kubelet[2823]: I0302 14:27:17.148336 2823 state_mem.go:36] "Initialized new in-memory state store" Mar 2 14:27:17.150130 kubelet[2823]: I0302 14:27:17.149726 2823 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 2 14:27:17.150130 kubelet[2823]: I0302 14:27:17.149744 2823 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 2 14:27:17.150130 kubelet[2823]: I0302 14:27:17.149766 2823 policy_none.go:49] "None policy: Start" Mar 2 14:27:17.150130 kubelet[2823]: I0302 14:27:17.149780 2823 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 2 14:27:17.150130 kubelet[2823]: I0302 14:27:17.149797 2823 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 2 14:27:17.150130 kubelet[2823]: I0302 14:27:17.149902 2823 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Mar 2 14:27:17.150130 kubelet[2823]: I0302 14:27:17.150028 2823 policy_none.go:47] "Start" Mar 2 14:27:17.193291 kubelet[2823]: E0302 14:27:17.192189 2823 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 14:27:17.196049 kubelet[2823]: I0302 14:27:17.194124 2823 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 2 14:27:17.196049 kubelet[2823]: I0302 14:27:17.194165 2823 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 14:27:17.196049 kubelet[2823]: I0302 14:27:17.194631 2823 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 2 14:27:17.197044 kubelet[2823]: E0302 14:27:17.197025 2823 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 14:27:17.206025 kubelet[2823]: I0302 14:27:17.205103 2823 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 2 14:27:17.206025 kubelet[2823]: I0302 14:27:17.205301 2823 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:17.206025 kubelet[2823]: I0302 14:27:17.205819 2823 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:17.353419 kubelet[2823]: E0302 14:27:17.351763 2823 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:17.356902 kubelet[2823]: E0302 14:27:17.354343 2823 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Mar 2 14:27:17.356902 kubelet[2823]: I0302 14:27:17.355631 2823 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 14:27:17.366335 kubelet[2823]: I0302 14:27:17.365698 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:17.366335 kubelet[2823]: I0302 14:27:17.365744 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:17.366335 kubelet[2823]: I0302 14:27:17.365779 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/89efda49e166906783d8d868d41ebb86-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"89efda49e166906783d8d868d41ebb86\") " pod="kube-system/kube-scheduler-localhost" Mar 2 14:27:17.366335 kubelet[2823]: I0302 14:27:17.365806 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a8d35cbde6c3a030da03a196ce305505-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"a8d35cbde6c3a030da03a196ce305505\") " pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:17.366335 kubelet[2823]: I0302 14:27:17.365836 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:17.366655 kubelet[2823]: I0302 14:27:17.366192 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:17.366655 kubelet[2823]: I0302 14:27:17.366222 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 14:27:17.366655 kubelet[2823]: I0302 14:27:17.366246 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a8d35cbde6c3a030da03a196ce305505-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"a8d35cbde6c3a030da03a196ce305505\") " pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:17.366655 kubelet[2823]: I0302 14:27:17.366273 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a8d35cbde6c3a030da03a196ce305505-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"a8d35cbde6c3a030da03a196ce305505\") " pod="kube-system/kube-apiserver-localhost" Mar 2 14:27:17.427814 kubelet[2823]: I0302 14:27:17.424334 2823 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Mar 2 14:27:17.427814 kubelet[2823]: I0302 14:27:17.424498 2823 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Mar 2 14:27:17.569268 kubelet[2823]: E0302 14:27:17.567095 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:17.655132 kubelet[2823]: E0302 14:27:17.654279 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:17.655132 kubelet[2823]: E0302 14:27:17.654728 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:17.840271 kubelet[2823]: I0302 14:27:17.835159 2823 apiserver.go:52] "Watching apiserver" Mar 2 14:27:17.863498 kubelet[2823]: I0302 14:27:17.861513 2823 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 2 14:27:18.125529 kubelet[2823]: E0302 14:27:18.125378 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:18.126319 kubelet[2823]: E0302 14:27:18.126231 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:18.126668 kubelet[2823]: E0302 14:27:18.126649 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:18.243599 kubelet[2823]: I0302 14:27:18.238895 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.238874101 podStartE2EDuration="1.238874101s" podCreationTimestamp="2026-03-02 14:27:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 14:27:18.156406362 +0000 UTC m=+1.590074435" watchObservedRunningTime="2026-03-02 14:27:18.238874101 +0000 UTC m=+1.672542124" Mar 2 14:27:18.390048 kubelet[2823]: I0302 14:27:18.389011 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=7.388972244 podStartE2EDuration="7.388972244s" podCreationTimestamp="2026-03-02 14:27:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 14:27:18.310801184 +0000 UTC m=+1.744469186" watchObservedRunningTime="2026-03-02 14:27:18.388972244 +0000 UTC m=+1.822640247" Mar 2 14:27:19.134780 kubelet[2823]: E0302 14:27:19.134710 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:19.137262 kubelet[2823]: E0302 14:27:19.135378 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:19.950587 kubelet[2823]: I0302 14:27:19.950022 2823 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 2 14:27:19.951503 containerd[1574]: time="2026-03-02T14:27:19.951265000Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 2 14:27:19.954428 kubelet[2823]: I0302 14:27:19.954181 2823 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 2 14:27:20.139803 kubelet[2823]: E0302 14:27:20.139670 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:20.239563 kubelet[2823]: E0302 14:27:20.239285 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:20.762110 systemd[1]: Created slice kubepods-besteffort-pod6682b5c3_68b4_4421_a5d0_7562b4e27ea6.slice - libcontainer container kubepods-besteffort-pod6682b5c3_68b4_4421_a5d0_7562b4e27ea6.slice. Mar 2 14:27:20.829081 kubelet[2823]: I0302 14:27:20.827886 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6682b5c3-68b4-4421-a5d0-7562b4e27ea6-xtables-lock\") pod \"kube-proxy-zxwr8\" (UID: \"6682b5c3-68b4-4421-a5d0-7562b4e27ea6\") " pod="kube-system/kube-proxy-zxwr8" Mar 2 14:27:20.829081 kubelet[2823]: I0302 14:27:20.828041 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmcgv\" (UniqueName: \"kubernetes.io/projected/6682b5c3-68b4-4421-a5d0-7562b4e27ea6-kube-api-access-cmcgv\") pod \"kube-proxy-zxwr8\" (UID: \"6682b5c3-68b4-4421-a5d0-7562b4e27ea6\") " pod="kube-system/kube-proxy-zxwr8" Mar 2 14:27:20.829081 kubelet[2823]: I0302 14:27:20.828070 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/6682b5c3-68b4-4421-a5d0-7562b4e27ea6-kube-proxy\") pod \"kube-proxy-zxwr8\" (UID: \"6682b5c3-68b4-4421-a5d0-7562b4e27ea6\") " pod="kube-system/kube-proxy-zxwr8" Mar 2 14:27:20.829081 kubelet[2823]: I0302 14:27:20.828088 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6682b5c3-68b4-4421-a5d0-7562b4e27ea6-lib-modules\") pod \"kube-proxy-zxwr8\" (UID: \"6682b5c3-68b4-4421-a5d0-7562b4e27ea6\") " pod="kube-system/kube-proxy-zxwr8" Mar 2 14:27:21.105066 kubelet[2823]: E0302 14:27:21.104060 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:21.113716 containerd[1574]: time="2026-03-02T14:27:21.113265919Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-zxwr8,Uid:6682b5c3-68b4-4421-a5d0-7562b4e27ea6,Namespace:kube-system,Attempt:0,}" Mar 2 14:27:21.141681 kubelet[2823]: E0302 14:27:21.141646 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:21.214095 containerd[1574]: time="2026-03-02T14:27:21.213356050Z" level=info msg="connecting to shim f59e24a7603223a062474f238d0a467dc0cea35402d78bbd96eb1649aa3461e2" address="unix:///run/containerd/s/31e1b730f805ade93f73984dc6e3ca286e0e9f226c519e7a66cbff989944b310" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:27:21.291762 systemd[1]: Created slice kubepods-besteffort-pod389b7152_4d33_4357_92e3_42d63f7bb95d.slice - libcontainer container kubepods-besteffort-pod389b7152_4d33_4357_92e3_42d63f7bb95d.slice. Mar 2 14:27:21.321792 systemd[1]: Started cri-containerd-f59e24a7603223a062474f238d0a467dc0cea35402d78bbd96eb1649aa3461e2.scope - libcontainer container f59e24a7603223a062474f238d0a467dc0cea35402d78bbd96eb1649aa3461e2. Mar 2 14:27:21.332217 kubelet[2823]: I0302 14:27:21.331771 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cblcs\" (UniqueName: \"kubernetes.io/projected/389b7152-4d33-4357-92e3-42d63f7bb95d-kube-api-access-cblcs\") pod \"tigera-operator-85979684d8-s8vkd\" (UID: \"389b7152-4d33-4357-92e3-42d63f7bb95d\") " pod="tigera-operator/tigera-operator-85979684d8-s8vkd" Mar 2 14:27:21.332217 kubelet[2823]: I0302 14:27:21.331996 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/389b7152-4d33-4357-92e3-42d63f7bb95d-var-lib-calico\") pod \"tigera-operator-85979684d8-s8vkd\" (UID: \"389b7152-4d33-4357-92e3-42d63f7bb95d\") " pod="tigera-operator/tigera-operator-85979684d8-s8vkd" Mar 2 14:27:21.496799 containerd[1574]: time="2026-03-02T14:27:21.494544814Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-zxwr8,Uid:6682b5c3-68b4-4421-a5d0-7562b4e27ea6,Namespace:kube-system,Attempt:0,} returns sandbox id \"f59e24a7603223a062474f238d0a467dc0cea35402d78bbd96eb1649aa3461e2\"" Mar 2 14:27:21.506417 kubelet[2823]: E0302 14:27:21.506302 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:21.519298 containerd[1574]: time="2026-03-02T14:27:21.517171672Z" level=info msg="CreateContainer within sandbox \"f59e24a7603223a062474f238d0a467dc0cea35402d78bbd96eb1649aa3461e2\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 2 14:27:21.552827 containerd[1574]: time="2026-03-02T14:27:21.552671099Z" level=info msg="Container 8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:27:21.584701 containerd[1574]: time="2026-03-02T14:27:21.583637767Z" level=info msg="CreateContainer within sandbox \"f59e24a7603223a062474f238d0a467dc0cea35402d78bbd96eb1649aa3461e2\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43\"" Mar 2 14:27:21.586867 containerd[1574]: time="2026-03-02T14:27:21.586666269Z" level=info msg="StartContainer for \"8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43\"" Mar 2 14:27:21.591636 containerd[1574]: time="2026-03-02T14:27:21.591371620Z" level=info msg="connecting to shim 8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43" address="unix:///run/containerd/s/31e1b730f805ade93f73984dc6e3ca286e0e9f226c519e7a66cbff989944b310" protocol=ttrpc version=3 Mar 2 14:27:21.613509 containerd[1574]: time="2026-03-02T14:27:21.610832174Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-85979684d8-s8vkd,Uid:389b7152-4d33-4357-92e3-42d63f7bb95d,Namespace:tigera-operator,Attempt:0,}" Mar 2 14:27:21.653811 systemd[1]: Started cri-containerd-8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43.scope - libcontainer container 8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43. Mar 2 14:27:21.694513 containerd[1574]: time="2026-03-02T14:27:21.694394662Z" level=info msg="connecting to shim d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7" address="unix:///run/containerd/s/d50d4af1d30650ac9f94f6a45da621e80f2619c1b270b079a50b9ac49ce8551d" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:27:21.766810 systemd[1]: Started cri-containerd-d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7.scope - libcontainer container d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7. Mar 2 14:27:21.969140 containerd[1574]: time="2026-03-02T14:27:21.965226786Z" level=info msg="StartContainer for \"8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43\" returns successfully" Mar 2 14:27:22.027622 containerd[1574]: time="2026-03-02T14:27:22.027408958Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-85979684d8-s8vkd,Uid:389b7152-4d33-4357-92e3-42d63f7bb95d,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7\"" Mar 2 14:27:22.035239 containerd[1574]: time="2026-03-02T14:27:22.035129692Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.3\"" Mar 2 14:27:22.166779 kubelet[2823]: E0302 14:27:22.166271 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:22.166779 kubelet[2823]: E0302 14:27:22.166628 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:22.814229 kubelet[2823]: E0302 14:27:22.812755 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:22.891064 kubelet[2823]: I0302 14:27:22.889519 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-zxwr8" podStartSLOduration=2.889497066 podStartE2EDuration="2.889497066s" podCreationTimestamp="2026-03-02 14:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 14:27:22.236520333 +0000 UTC m=+5.670188357" watchObservedRunningTime="2026-03-02 14:27:22.889497066 +0000 UTC m=+6.323165098" Mar 2 14:27:23.173044 kubelet[2823]: E0302 14:27:23.172779 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:23.274735 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3664295785.mount: Deactivated successfully. Mar 2 14:27:27.715642 containerd[1574]: time="2026-03-02T14:27:27.715321338Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:27:27.717815 containerd[1574]: time="2026-03-02T14:27:27.717052660Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.3: active requests=0, bytes read=40822719" Mar 2 14:27:27.719130 containerd[1574]: time="2026-03-02T14:27:27.718830339Z" level=info msg="ImageCreate event name:\"sha256:de15454df5913bb69360783a4d76287caf2c87324eed18162e79d4c06a4c8896\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:27:27.732911 containerd[1574]: time="2026-03-02T14:27:27.732751659Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:3b1a6762e1f3fae8490773b8f06ddd1e6775850febbece4d6002416f39adc670\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:27:27.738274 containerd[1574]: time="2026-03-02T14:27:27.738071551Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.3\" with image id \"sha256:de15454df5913bb69360783a4d76287caf2c87324eed18162e79d4c06a4c8896\", repo tag \"quay.io/tigera/operator:v1.40.3\", repo digest \"quay.io/tigera/operator@sha256:3b1a6762e1f3fae8490773b8f06ddd1e6775850febbece4d6002416f39adc670\", size \"40818714\" in 5.702652589s" Mar 2 14:27:27.738274 containerd[1574]: time="2026-03-02T14:27:27.738161429Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.3\" returns image reference \"sha256:de15454df5913bb69360783a4d76287caf2c87324eed18162e79d4c06a4c8896\"" Mar 2 14:27:27.752650 containerd[1574]: time="2026-03-02T14:27:27.752141890Z" level=info msg="CreateContainer within sandbox \"d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Mar 2 14:27:27.788812 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount445941401.mount: Deactivated successfully. Mar 2 14:27:27.793329 containerd[1574]: time="2026-03-02T14:27:27.789798060Z" level=info msg="Container 0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:27:27.819259 containerd[1574]: time="2026-03-02T14:27:27.819127055Z" level=info msg="CreateContainer within sandbox \"d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a\"" Mar 2 14:27:27.820915 containerd[1574]: time="2026-03-02T14:27:27.820767685Z" level=info msg="StartContainer for \"0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a\"" Mar 2 14:27:27.824688 containerd[1574]: time="2026-03-02T14:27:27.824110666Z" level=info msg="connecting to shim 0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a" address="unix:///run/containerd/s/d50d4af1d30650ac9f94f6a45da621e80f2619c1b270b079a50b9ac49ce8551d" protocol=ttrpc version=3 Mar 2 14:27:27.868696 systemd[1]: Started cri-containerd-0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a.scope - libcontainer container 0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a. Mar 2 14:27:27.951498 containerd[1574]: time="2026-03-02T14:27:27.949018285Z" level=info msg="StartContainer for \"0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a\" returns successfully" Mar 2 14:27:28.259762 kubelet[2823]: I0302 14:27:28.258279 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-85979684d8-s8vkd" podStartSLOduration=1.545649356 podStartE2EDuration="7.258262781s" podCreationTimestamp="2026-03-02 14:27:21 +0000 UTC" firstStartedPulling="2026-03-02 14:27:22.030215528 +0000 UTC m=+5.463883531" lastFinishedPulling="2026-03-02 14:27:27.742828953 +0000 UTC m=+11.176496956" observedRunningTime="2026-03-02 14:27:28.256237398 +0000 UTC m=+11.689905411" watchObservedRunningTime="2026-03-02 14:27:28.258262781 +0000 UTC m=+11.691930784" Mar 2 14:27:32.387882 systemd[1]: cri-containerd-0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a.scope: Deactivated successfully. Mar 2 14:27:32.394919 containerd[1574]: time="2026-03-02T14:27:32.394811891Z" level=info msg="received container exit event container_id:\"0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a\" id:\"0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a\" pid:3158 exit_status:1 exited_at:{seconds:1772461652 nanos:393881977}" Mar 2 14:27:32.622587 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a-rootfs.mount: Deactivated successfully. Mar 2 14:27:33.252773 kubelet[2823]: I0302 14:27:33.252699 2823 scope.go:117] "RemoveContainer" containerID="0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a" Mar 2 14:27:33.265950 containerd[1574]: time="2026-03-02T14:27:33.265861161Z" level=info msg="CreateContainer within sandbox \"d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Mar 2 14:27:33.343069 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2906768048.mount: Deactivated successfully. Mar 2 14:27:33.387271 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2445891656.mount: Deactivated successfully. Mar 2 14:27:33.398699 containerd[1574]: time="2026-03-02T14:27:33.396329296Z" level=info msg="Container 0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:27:33.439500 containerd[1574]: time="2026-03-02T14:27:33.439346161Z" level=info msg="CreateContainer within sandbox \"d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812\"" Mar 2 14:27:33.447288 containerd[1574]: time="2026-03-02T14:27:33.447008686Z" level=info msg="StartContainer for \"0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812\"" Mar 2 14:27:33.452292 containerd[1574]: time="2026-03-02T14:27:33.452200390Z" level=info msg="connecting to shim 0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812" address="unix:///run/containerd/s/d50d4af1d30650ac9f94f6a45da621e80f2619c1b270b079a50b9ac49ce8551d" protocol=ttrpc version=3 Mar 2 14:27:33.509357 systemd[1]: Started cri-containerd-0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812.scope - libcontainer container 0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812. Mar 2 14:27:33.665851 containerd[1574]: time="2026-03-02T14:27:33.665769703Z" level=info msg="StartContainer for \"0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812\" returns successfully" Mar 2 14:27:37.268528 sudo[1807]: pam_unix(sudo:session): session closed for user root Mar 2 14:27:37.280379 sshd[1806]: Connection closed by 10.0.0.1 port 33364 Mar 2 14:27:37.283348 sshd-session[1801]: pam_unix(sshd:session): session closed for user core Mar 2 14:27:37.297397 systemd[1]: sshd@8-10.0.0.9:22-10.0.0.1:33364.service: Deactivated successfully. Mar 2 14:27:37.311205 systemd[1]: session-9.scope: Deactivated successfully. Mar 2 14:27:37.313317 systemd[1]: session-9.scope: Consumed 12.133s CPU time, 231.6M memory peak. Mar 2 14:27:37.321978 systemd-logind[1554]: Session 9 logged out. Waiting for processes to exit. Mar 2 14:27:37.330264 systemd-logind[1554]: Removed session 9. Mar 2 14:27:46.472855 systemd[1]: Created slice kubepods-besteffort-pod404f3cd9_5a00_48bc_b7c9_ef7413827faf.slice - libcontainer container kubepods-besteffort-pod404f3cd9_5a00_48bc_b7c9_ef7413827faf.slice. Mar 2 14:27:46.503066 kubelet[2823]: I0302 14:27:46.500683 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/404f3cd9-5a00-48bc-b7c9-ef7413827faf-tigera-ca-bundle\") pod \"calico-typha-78dff98b8f-fh7br\" (UID: \"404f3cd9-5a00-48bc-b7c9-ef7413827faf\") " pod="calico-system/calico-typha-78dff98b8f-fh7br" Mar 2 14:27:46.503066 kubelet[2823]: I0302 14:27:46.500753 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/404f3cd9-5a00-48bc-b7c9-ef7413827faf-typha-certs\") pod \"calico-typha-78dff98b8f-fh7br\" (UID: \"404f3cd9-5a00-48bc-b7c9-ef7413827faf\") " pod="calico-system/calico-typha-78dff98b8f-fh7br" Mar 2 14:27:46.503066 kubelet[2823]: I0302 14:27:46.500780 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzqgj\" (UniqueName: \"kubernetes.io/projected/404f3cd9-5a00-48bc-b7c9-ef7413827faf-kube-api-access-vzqgj\") pod \"calico-typha-78dff98b8f-fh7br\" (UID: \"404f3cd9-5a00-48bc-b7c9-ef7413827faf\") " pod="calico-system/calico-typha-78dff98b8f-fh7br" Mar 2 14:27:47.121201 kubelet[2823]: E0302 14:27:47.121076 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:47.180089 containerd[1574]: time="2026-03-02T14:27:47.179815466Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-78dff98b8f-fh7br,Uid:404f3cd9-5a00-48bc-b7c9-ef7413827faf,Namespace:calico-system,Attempt:0,}" Mar 2 14:27:48.099417 kubelet[2823]: I0302 14:27:48.075039 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-node-certs\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.099417 kubelet[2823]: I0302 14:27:48.075564 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-var-run-calico\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.099417 kubelet[2823]: I0302 14:27:48.075635 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-sys-fs\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.099417 kubelet[2823]: I0302 14:27:48.075659 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-cni-log-dir\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.099417 kubelet[2823]: I0302 14:27:48.075684 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-cni-bin-dir\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200431 kubelet[2823]: I0302 14:27:48.075706 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-lib-modules\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200431 kubelet[2823]: I0302 14:27:48.075725 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-nodeproc\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200431 kubelet[2823]: I0302 14:27:48.075793 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-tigera-ca-bundle\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200431 kubelet[2823]: I0302 14:27:48.075812 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-xtables-lock\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200431 kubelet[2823]: I0302 14:27:48.075837 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwv5n\" (UniqueName: \"kubernetes.io/projected/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-kube-api-access-qwv5n\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200670 kubelet[2823]: I0302 14:27:48.075874 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-flexvol-driver-host\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200670 kubelet[2823]: I0302 14:27:48.075893 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-bpffs\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200670 kubelet[2823]: I0302 14:27:48.075918 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-cni-net-dir\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200670 kubelet[2823]: I0302 14:27:48.075940 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-policysync\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.200670 kubelet[2823]: I0302 14:27:48.076066 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a-var-lib-calico\") pod \"calico-node-mbhdq\" (UID: \"89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a\") " pod="calico-system/calico-node-mbhdq" Mar 2 14:27:48.280383 kubelet[2823]: E0302 14:27:48.279643 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.321484 kubelet[2823]: W0302 14:27:48.316878 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.328311 kubelet[2823]: E0302 14:27:48.322394 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.417120 kubelet[2823]: E0302 14:27:48.411675 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.417120 kubelet[2823]: W0302 14:27:48.411758 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.417120 kubelet[2823]: E0302 14:27:48.411794 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.417120 kubelet[2823]: E0302 14:27:48.416840 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.417120 kubelet[2823]: W0302 14:27:48.416864 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.417120 kubelet[2823]: E0302 14:27:48.416900 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.419984 kubelet[2823]: E0302 14:27:48.418580 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.419984 kubelet[2823]: W0302 14:27:48.418594 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.419984 kubelet[2823]: E0302 14:27:48.418609 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.419984 kubelet[2823]: E0302 14:27:48.419068 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.419984 kubelet[2823]: W0302 14:27:48.419081 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.419984 kubelet[2823]: E0302 14:27:48.419093 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.450314 kubelet[2823]: E0302 14:27:48.425919 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.450314 kubelet[2823]: W0302 14:27:48.426006 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.450314 kubelet[2823]: E0302 14:27:48.426046 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.456212 kubelet[2823]: E0302 14:27:48.455747 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.502241 kubelet[2823]: W0302 14:27:48.459860 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.502241 kubelet[2823]: E0302 14:27:48.460053 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.516149 kubelet[2823]: E0302 14:27:48.516083 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.517144 kubelet[2823]: W0302 14:27:48.517118 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.517426 kubelet[2823]: E0302 14:27:48.517234 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.533318 kubelet[2823]: E0302 14:27:48.532613 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.533318 kubelet[2823]: W0302 14:27:48.532651 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.533318 kubelet[2823]: E0302 14:27:48.532701 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.553431 kubelet[2823]: E0302 14:27:48.553038 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.553431 kubelet[2823]: W0302 14:27:48.553101 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.553431 kubelet[2823]: E0302 14:27:48.553156 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.601288 kubelet[2823]: E0302 14:27:48.601187 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.617871 kubelet[2823]: W0302 14:27:48.601555 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.617871 kubelet[2823]: E0302 14:27:48.601587 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.617871 kubelet[2823]: E0302 14:27:48.607159 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.617871 kubelet[2823]: W0302 14:27:48.607279 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.617871 kubelet[2823]: E0302 14:27:48.607332 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.703498 kubelet[2823]: E0302 14:27:48.652553 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.703498 kubelet[2823]: W0302 14:27:48.681736 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.703498 kubelet[2823]: E0302 14:27:48.700657 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.864620 systemd[1]: Created slice kubepods-besteffort-pod89a4e8cd_0f0e_4f2b_912c_3aceb3d7661a.slice - libcontainer container kubepods-besteffort-pod89a4e8cd_0f0e_4f2b_912c_3aceb3d7661a.slice. Mar 2 14:27:48.910932 kubelet[2823]: E0302 14:27:48.908496 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:27:48.947596 kubelet[2823]: E0302 14:27:48.946577 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.978895 kubelet[2823]: W0302 14:27:48.964385 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.978895 kubelet[2823]: E0302 14:27:48.967410 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.978895 kubelet[2823]: E0302 14:27:48.968262 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.978895 kubelet[2823]: W0302 14:27:48.968277 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.978895 kubelet[2823]: E0302 14:27:48.968294 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.978895 kubelet[2823]: E0302 14:27:48.968574 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.978895 kubelet[2823]: W0302 14:27:48.968587 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:48.978895 kubelet[2823]: E0302 14:27:48.968610 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:48.978895 kubelet[2823]: E0302 14:27:48.973239 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:48.978895 kubelet[2823]: W0302 14:27:48.973355 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.005583 kubelet[2823]: E0302 14:27:48.973395 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.005583 kubelet[2823]: E0302 14:27:48.974710 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.005583 kubelet[2823]: W0302 14:27:48.974725 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.005583 kubelet[2823]: E0302 14:27:48.974738 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.005583 kubelet[2823]: E0302 14:27:48.997136 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.005583 kubelet[2823]: W0302 14:27:48.997222 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.005583 kubelet[2823]: E0302 14:27:48.997267 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.005583 kubelet[2823]: E0302 14:27:49.001019 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.005583 kubelet[2823]: W0302 14:27:49.001037 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.005583 kubelet[2823]: E0302 14:27:49.001053 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.274850 kubelet[2823]: E0302 14:27:49.002569 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.274850 kubelet[2823]: W0302 14:27:49.002585 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.274850 kubelet[2823]: E0302 14:27:49.002597 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.274850 kubelet[2823]: E0302 14:27:49.004367 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.274850 kubelet[2823]: W0302 14:27:49.004380 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.274850 kubelet[2823]: E0302 14:27:49.004399 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.274850 kubelet[2823]: E0302 14:27:49.005512 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.274850 kubelet[2823]: W0302 14:27:49.005524 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.274850 kubelet[2823]: E0302 14:27:49.005535 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.274850 kubelet[2823]: E0302 14:27:49.006762 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.409706 kubelet[2823]: W0302 14:27:49.006774 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.409706 kubelet[2823]: E0302 14:27:49.006785 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.409706 kubelet[2823]: E0302 14:27:49.013802 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.409706 kubelet[2823]: W0302 14:27:49.013905 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.409706 kubelet[2823]: E0302 14:27:49.013981 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.409706 kubelet[2823]: E0302 14:27:49.037144 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.409706 kubelet[2823]: W0302 14:27:49.037283 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.409706 kubelet[2823]: E0302 14:27:49.037337 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.409706 kubelet[2823]: I0302 14:27:49.037668 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f81f0355-d756-4ec2-b9d0-9639b176b2ea-registration-dir\") pod \"csi-node-driver-fgxfk\" (UID: \"f81f0355-d756-4ec2-b9d0-9639b176b2ea\") " pod="calico-system/csi-node-driver-fgxfk" Mar 2 14:27:49.410589 kubelet[2823]: E0302 14:27:49.038113 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.410589 kubelet[2823]: W0302 14:27:49.038128 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.410589 kubelet[2823]: E0302 14:27:49.038142 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.410589 kubelet[2823]: E0302 14:27:49.038561 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.410589 kubelet[2823]: W0302 14:27:49.038571 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.410589 kubelet[2823]: E0302 14:27:49.038582 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.410589 kubelet[2823]: E0302 14:27:49.041532 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.410589 kubelet[2823]: W0302 14:27:49.041547 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.410589 kubelet[2823]: E0302 14:27:49.041560 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.410589 kubelet[2823]: E0302 14:27:49.042723 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.410998 kubelet[2823]: W0302 14:27:49.042734 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.410998 kubelet[2823]: E0302 14:27:49.042746 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.410998 kubelet[2823]: E0302 14:27:49.057398 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.410998 kubelet[2823]: W0302 14:27:49.057422 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.410998 kubelet[2823]: E0302 14:27:49.057550 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.410998 kubelet[2823]: I0302 14:27:49.058202 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f81f0355-d756-4ec2-b9d0-9639b176b2ea-socket-dir\") pod \"csi-node-driver-fgxfk\" (UID: \"f81f0355-d756-4ec2-b9d0-9639b176b2ea\") " pod="calico-system/csi-node-driver-fgxfk" Mar 2 14:27:49.410998 kubelet[2823]: E0302 14:27:49.075690 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.410998 kubelet[2823]: W0302 14:27:49.075770 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.410998 kubelet[2823]: E0302 14:27:49.075824 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.441168 kubelet[2823]: E0302 14:27:49.156393 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.441168 kubelet[2823]: W0302 14:27:49.156426 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.441168 kubelet[2823]: E0302 14:27:49.156535 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.441168 kubelet[2823]: E0302 14:27:49.237155 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.441168 kubelet[2823]: W0302 14:27:49.237223 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.441168 kubelet[2823]: E0302 14:27:49.237308 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.441168 kubelet[2823]: E0302 14:27:49.246493 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.441168 kubelet[2823]: W0302 14:27:49.246517 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.441168 kubelet[2823]: E0302 14:27:49.260766 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.441168 kubelet[2823]: E0302 14:27:49.299911 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.441863 kubelet[2823]: W0302 14:27:49.303131 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.441863 kubelet[2823]: E0302 14:27:49.303182 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.441863 kubelet[2823]: E0302 14:27:49.398575 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.441863 kubelet[2823]: W0302 14:27:49.398675 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.441863 kubelet[2823]: E0302 14:27:49.398720 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.441863 kubelet[2823]: I0302 14:27:49.399308 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f81f0355-d756-4ec2-b9d0-9639b176b2ea-kubelet-dir\") pod \"csi-node-driver-fgxfk\" (UID: \"f81f0355-d756-4ec2-b9d0-9639b176b2ea\") " pod="calico-system/csi-node-driver-fgxfk" Mar 2 14:27:49.441863 kubelet[2823]: E0302 14:27:49.409639 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.441863 kubelet[2823]: W0302 14:27:49.409672 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.441863 kubelet[2823]: E0302 14:27:49.409717 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.442176 kubelet[2823]: E0302 14:27:49.410382 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.442176 kubelet[2823]: W0302 14:27:49.410396 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.442176 kubelet[2823]: E0302 14:27:49.410414 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.442176 kubelet[2823]: E0302 14:27:49.410900 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.442176 kubelet[2823]: W0302 14:27:49.410914 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.442176 kubelet[2823]: E0302 14:27:49.410926 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.442176 kubelet[2823]: E0302 14:27:49.411248 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.442176 kubelet[2823]: W0302 14:27:49.411263 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.442176 kubelet[2823]: E0302 14:27:49.411294 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.442176 kubelet[2823]: E0302 14:27:49.411676 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.446123 kubelet[2823]: W0302 14:27:49.411688 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.446123 kubelet[2823]: E0302 14:27:49.411706 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.607832 kubelet[2823]: E0302 14:27:49.604077 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.607832 kubelet[2823]: W0302 14:27:49.604547 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.607832 kubelet[2823]: E0302 14:27:49.604651 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.608389 containerd[1574]: time="2026-03-02T14:27:49.605714843Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-mbhdq,Uid:89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a,Namespace:calico-system,Attempt:0,}" Mar 2 14:27:49.615985 kubelet[2823]: E0302 14:27:49.615806 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.615985 kubelet[2823]: W0302 14:27:49.615839 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.615985 kubelet[2823]: E0302 14:27:49.615867 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.638554 kubelet[2823]: E0302 14:27:49.638335 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.666693 kubelet[2823]: W0302 14:27:49.638430 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.666693 kubelet[2823]: E0302 14:27:49.665574 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.700852 kubelet[2823]: E0302 14:27:49.700130 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.714421 kubelet[2823]: W0302 14:27:49.712236 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.762561 kubelet[2823]: E0302 14:27:49.719596 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.762561 kubelet[2823]: E0302 14:27:49.755803 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.762561 kubelet[2823]: W0302 14:27:49.761168 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.762561 kubelet[2823]: E0302 14:27:49.761498 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:49.969403 kubelet[2823]: E0302 14:27:49.967251 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:49.969403 kubelet[2823]: W0302 14:27:49.968853 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:49.969403 kubelet[2823]: E0302 14:27:49.969146 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.074099 containerd[1574]: time="2026-03-02T14:27:50.070956799Z" level=info msg="connecting to shim c25234fbbf3c026a1878d06769dbd0e187e30eb05bca8798541087348e8b5c45" address="unix:///run/containerd/s/cd7c9448a7c15dbb4e19d27d729d5020e87b8f38c3d1ba02e9734cf48bb19525" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:27:50.112313 kubelet[2823]: E0302 14:27:50.082117 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.112313 kubelet[2823]: W0302 14:27:50.082339 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.141501 kubelet[2823]: E0302 14:27:50.133222 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.156663 kubelet[2823]: E0302 14:27:50.156229 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.157487 kubelet[2823]: W0302 14:27:50.157233 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.166924 kubelet[2823]: E0302 14:27:50.159557 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.199327 kubelet[2823]: E0302 14:27:50.198592 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.199327 kubelet[2823]: W0302 14:27:50.200818 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.199327 kubelet[2823]: E0302 14:27:50.201055 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.218520 kubelet[2823]: E0302 14:27:50.217922 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.218520 kubelet[2823]: W0302 14:27:50.217987 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.218520 kubelet[2823]: E0302 14:27:50.218024 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.265718 kubelet[2823]: E0302 14:27:50.221703 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.265718 kubelet[2823]: W0302 14:27:50.221724 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.265718 kubelet[2823]: E0302 14:27:50.221745 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.305848 kubelet[2823]: E0302 14:27:50.305667 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.305848 kubelet[2823]: W0302 14:27:50.305793 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.305848 kubelet[2823]: E0302 14:27:50.305845 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.313628 kubelet[2823]: E0302 14:27:50.313602 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.313628 kubelet[2823]: W0302 14:27:50.313624 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.313755 kubelet[2823]: E0302 14:27:50.313645 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.378709 kubelet[2823]: E0302 14:27:50.378640 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.378709 kubelet[2823]: W0302 14:27:50.378694 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.379058 kubelet[2823]: E0302 14:27:50.378730 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.407797 kubelet[2823]: E0302 14:27:50.407676 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.407797 kubelet[2823]: W0302 14:27:50.407717 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.416114 kubelet[2823]: E0302 14:27:50.415968 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.419866 kubelet[2823]: E0302 14:27:50.419606 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.419866 kubelet[2823]: W0302 14:27:50.419651 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.419866 kubelet[2823]: E0302 14:27:50.419675 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.422093 kubelet[2823]: E0302 14:27:50.422073 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.422192 kubelet[2823]: W0302 14:27:50.422177 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.422264 kubelet[2823]: E0302 14:27:50.422251 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.422674 kubelet[2823]: E0302 14:27:50.422656 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.422769 kubelet[2823]: W0302 14:27:50.422752 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.422860 kubelet[2823]: E0302 14:27:50.422842 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.423726 kubelet[2823]: E0302 14:27:50.423711 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.423810 kubelet[2823]: W0302 14:27:50.423797 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.423882 kubelet[2823]: E0302 14:27:50.423869 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.424263 kubelet[2823]: E0302 14:27:50.424217 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.424356 kubelet[2823]: W0302 14:27:50.424339 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.424517 kubelet[2823]: E0302 14:27:50.424497 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.424860 kubelet[2823]: E0302 14:27:50.424843 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.426305 kubelet[2823]: W0302 14:27:50.426285 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.426413 kubelet[2823]: E0302 14:27:50.426394 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.427739 kubelet[2823]: E0302 14:27:50.427723 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.427825 kubelet[2823]: W0302 14:27:50.427812 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.427901 kubelet[2823]: E0302 14:27:50.427887 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.428997 kubelet[2823]: E0302 14:27:50.428978 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.429079 kubelet[2823]: W0302 14:27:50.429061 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.429172 kubelet[2823]: E0302 14:27:50.429157 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.433068 kubelet[2823]: E0302 14:27:50.433050 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.433549 kubelet[2823]: W0302 14:27:50.433209 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.433549 kubelet[2823]: E0302 14:27:50.433308 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.433889 kubelet[2823]: E0302 14:27:50.433824 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.433889 kubelet[2823]: W0302 14:27:50.433871 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.433889 kubelet[2823]: E0302 14:27:50.433886 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.434625 kubelet[2823]: E0302 14:27:50.434350 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.434625 kubelet[2823]: W0302 14:27:50.434368 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.434625 kubelet[2823]: E0302 14:27:50.434380 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.437148 kubelet[2823]: E0302 14:27:50.436902 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.437148 kubelet[2823]: W0302 14:27:50.437033 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.437148 kubelet[2823]: E0302 14:27:50.437050 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.442415 kubelet[2823]: E0302 14:27:50.442300 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.442415 kubelet[2823]: W0302 14:27:50.442317 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.442415 kubelet[2823]: E0302 14:27:50.442333 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.445699 kubelet[2823]: E0302 14:27:50.445648 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.446049 kubelet[2823]: W0302 14:27:50.445817 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.446049 kubelet[2823]: E0302 14:27:50.445835 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.449125 kubelet[2823]: E0302 14:27:50.447479 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.449745 kubelet[2823]: W0302 14:27:50.449502 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.449745 kubelet[2823]: E0302 14:27:50.449527 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.450537 kubelet[2823]: E0302 14:27:50.450520 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.450769 kubelet[2823]: W0302 14:27:50.450723 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.451187 kubelet[2823]: E0302 14:27:50.451168 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.454603 kubelet[2823]: E0302 14:27:50.454586 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.455524 kubelet[2823]: W0302 14:27:50.455502 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.455764 kubelet[2823]: E0302 14:27:50.455743 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.457154 containerd[1574]: time="2026-03-02T14:27:50.456727150Z" level=info msg="connecting to shim 5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b" address="unix:///run/containerd/s/0b9e2f1b1317597f174e05833aa8896ff5dba404179e70ab1652b5f1b44c0e65" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:27:50.459848 kubelet[2823]: E0302 14:27:50.459582 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.459848 kubelet[2823]: W0302 14:27:50.459597 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.459848 kubelet[2823]: E0302 14:27:50.459611 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.462465 kubelet[2823]: E0302 14:27:50.460974 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.462774 kubelet[2823]: W0302 14:27:50.462751 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.463621 kubelet[2823]: E0302 14:27:50.462966 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.464120 kubelet[2823]: E0302 14:27:50.464105 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.464389 kubelet[2823]: W0302 14:27:50.464353 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.464567 kubelet[2823]: E0302 14:27:50.464551 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.466594 kubelet[2823]: E0302 14:27:50.466578 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.467679 kubelet[2823]: W0302 14:27:50.467527 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.467679 kubelet[2823]: E0302 14:27:50.467552 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.469047 kubelet[2823]: E0302 14:27:50.468713 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.469241 kubelet[2823]: W0302 14:27:50.469224 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.469319 kubelet[2823]: E0302 14:27:50.469307 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.470592 kubelet[2823]: E0302 14:27:50.470338 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.470592 kubelet[2823]: W0302 14:27:50.470428 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.470592 kubelet[2823]: E0302 14:27:50.470488 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.471866 kubelet[2823]: E0302 14:27:50.471849 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.472139 kubelet[2823]: W0302 14:27:50.471983 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.472139 kubelet[2823]: E0302 14:27:50.472003 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.472626 kubelet[2823]: E0302 14:27:50.472611 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.473219 kubelet[2823]: W0302 14:27:50.472880 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.473219 kubelet[2823]: E0302 14:27:50.472898 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.475050 kubelet[2823]: E0302 14:27:50.475034 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.475352 kubelet[2823]: W0302 14:27:50.475182 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.475352 kubelet[2823]: E0302 14:27:50.475204 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.475713 kubelet[2823]: E0302 14:27:50.475566 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.475713 kubelet[2823]: W0302 14:27:50.475579 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.475713 kubelet[2823]: E0302 14:27:50.475590 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.476341 kubelet[2823]: E0302 14:27:50.476202 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.476341 kubelet[2823]: W0302 14:27:50.476218 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.476341 kubelet[2823]: E0302 14:27:50.476232 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.476872 kubelet[2823]: E0302 14:27:50.476859 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.477105 kubelet[2823]: W0302 14:27:50.477017 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.477105 kubelet[2823]: E0302 14:27:50.477035 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.478034 kubelet[2823]: E0302 14:27:50.478014 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.478224 kubelet[2823]: W0302 14:27:50.478205 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.478297 kubelet[2823]: E0302 14:27:50.478284 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.478742 kubelet[2823]: E0302 14:27:50.478724 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.478841 kubelet[2823]: W0302 14:27:50.478824 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.487706 kubelet[2823]: E0302 14:27:50.478910 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.499870 kubelet[2823]: E0302 14:27:50.499562 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.500798 kubelet[2823]: W0302 14:27:50.500042 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.500798 kubelet[2823]: E0302 14:27:50.500069 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.508870 kubelet[2823]: E0302 14:27:50.505054 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.513958 kubelet[2823]: W0302 14:27:50.511173 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.513958 kubelet[2823]: E0302 14:27:50.511402 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.513958 kubelet[2823]: E0302 14:27:50.513183 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.513958 kubelet[2823]: W0302 14:27:50.513710 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.513958 kubelet[2823]: E0302 14:27:50.513730 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.515431 kubelet[2823]: E0302 14:27:50.515412 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.515431 kubelet[2823]: W0302 14:27:50.515659 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.515431 kubelet[2823]: E0302 14:27:50.515678 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.521359 kubelet[2823]: E0302 14:27:50.521243 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.521359 kubelet[2823]: W0302 14:27:50.521261 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.521359 kubelet[2823]: E0302 14:27:50.521280 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.522719 kubelet[2823]: E0302 14:27:50.522554 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.522719 kubelet[2823]: W0302 14:27:50.522571 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.522719 kubelet[2823]: E0302 14:27:50.522601 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.528233 kubelet[2823]: E0302 14:27:50.527651 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.528233 kubelet[2823]: W0302 14:27:50.527674 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.528233 kubelet[2823]: E0302 14:27:50.527693 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.530168 kubelet[2823]: E0302 14:27:50.529622 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.530168 kubelet[2823]: W0302 14:27:50.529824 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.530168 kubelet[2823]: E0302 14:27:50.529839 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.536483 kubelet[2823]: E0302 14:27:50.535337 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.536483 kubelet[2823]: W0302 14:27:50.535393 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.536483 kubelet[2823]: E0302 14:27:50.535413 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.536794 kubelet[2823]: E0302 14:27:50.536776 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.537212 kubelet[2823]: W0302 14:27:50.536877 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.537308 kubelet[2823]: E0302 14:27:50.537291 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.542870 kubelet[2823]: E0302 14:27:50.542693 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.542870 kubelet[2823]: W0302 14:27:50.542715 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.542870 kubelet[2823]: E0302 14:27:50.542732 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.547674 kubelet[2823]: E0302 14:27:50.547606 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.547674 kubelet[2823]: W0302 14:27:50.547656 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.547674 kubelet[2823]: E0302 14:27:50.547674 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.557842 kubelet[2823]: E0302 14:27:50.548105 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.557842 kubelet[2823]: W0302 14:27:50.548323 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.557842 kubelet[2823]: E0302 14:27:50.548341 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.557842 kubelet[2823]: E0302 14:27:50.553545 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.557842 kubelet[2823]: W0302 14:27:50.553561 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.557842 kubelet[2823]: E0302 14:27:50.553656 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.565383 kubelet[2823]: E0302 14:27:50.565300 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.565383 kubelet[2823]: W0302 14:27:50.565330 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.565383 kubelet[2823]: E0302 14:27:50.565355 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.572108 kubelet[2823]: E0302 14:27:50.570311 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.572108 kubelet[2823]: W0302 14:27:50.570334 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.572108 kubelet[2823]: E0302 14:27:50.570358 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.572613 kubelet[2823]: E0302 14:27:50.572593 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.572613 kubelet[2823]: W0302 14:27:50.572612 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.572707 kubelet[2823]: E0302 14:27:50.572628 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.574730 systemd[1]: Started cri-containerd-c25234fbbf3c026a1878d06769dbd0e187e30eb05bca8798541087348e8b5c45.scope - libcontainer container c25234fbbf3c026a1878d06769dbd0e187e30eb05bca8798541087348e8b5c45. Mar 2 14:27:50.577726 kubelet[2823]: E0302 14:27:50.577544 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.577726 kubelet[2823]: W0302 14:27:50.577560 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.577726 kubelet[2823]: E0302 14:27:50.577578 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.586346 kubelet[2823]: E0302 14:27:50.585842 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.586346 kubelet[2823]: W0302 14:27:50.586071 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.586346 kubelet[2823]: E0302 14:27:50.586097 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.597230 kubelet[2823]: E0302 14:27:50.597164 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.597486 kubelet[2823]: W0302 14:27:50.597254 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.597486 kubelet[2823]: E0302 14:27:50.597279 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.600550 kubelet[2823]: E0302 14:27:50.598062 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.600550 kubelet[2823]: W0302 14:27:50.598148 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.600550 kubelet[2823]: E0302 14:27:50.598169 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.603028 kubelet[2823]: E0302 14:27:50.603005 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.603272 kubelet[2823]: W0302 14:27:50.603255 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.603540 kubelet[2823]: E0302 14:27:50.603520 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.607139 kubelet[2823]: E0302 14:27:50.607118 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.607226 kubelet[2823]: W0302 14:27:50.607212 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.607321 kubelet[2823]: E0302 14:27:50.607306 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.610799 kubelet[2823]: E0302 14:27:50.610537 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.610911 kubelet[2823]: W0302 14:27:50.610892 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.611034 kubelet[2823]: E0302 14:27:50.611018 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.611751 kubelet[2823]: E0302 14:27:50.611737 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.615847 kubelet[2823]: W0302 14:27:50.611860 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.616845 kubelet[2823]: E0302 14:27:50.616824 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.617312 kubelet[2823]: E0302 14:27:50.617294 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.617500 kubelet[2823]: W0302 14:27:50.617426 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.617592 kubelet[2823]: E0302 14:27:50.617575 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.621166 kubelet[2823]: E0302 14:27:50.621146 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.621280 kubelet[2823]: W0302 14:27:50.621261 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.621382 kubelet[2823]: E0302 14:27:50.621365 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.624617 kubelet[2823]: E0302 14:27:50.624541 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.624839 kubelet[2823]: W0302 14:27:50.624762 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.627182 kubelet[2823]: E0302 14:27:50.627096 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.635784 kubelet[2823]: E0302 14:27:50.635750 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.638269 kubelet[2823]: W0302 14:27:50.638245 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.638791 kubelet[2823]: E0302 14:27:50.638769 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.644636 kubelet[2823]: E0302 14:27:50.644536 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.645178 kubelet[2823]: W0302 14:27:50.645077 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.646033 kubelet[2823]: E0302 14:27:50.645430 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.646369 kubelet[2823]: E0302 14:27:50.646305 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.646551 kubelet[2823]: W0302 14:27:50.646532 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.646856 kubelet[2823]: E0302 14:27:50.646833 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.650856 kubelet[2823]: E0302 14:27:50.649543 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:27:50.660657 kubelet[2823]: E0302 14:27:50.660623 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.660799 kubelet[2823]: W0302 14:27:50.660777 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.661044 kubelet[2823]: E0302 14:27:50.660976 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.672671 kubelet[2823]: E0302 14:27:50.672639 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.672848 kubelet[2823]: W0302 14:27:50.672822 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.673002 kubelet[2823]: E0302 14:27:50.672978 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.675119 kubelet[2823]: E0302 14:27:50.675099 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.675228 kubelet[2823]: W0302 14:27:50.675212 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.675323 kubelet[2823]: E0302 14:27:50.675308 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.679216 kubelet[2823]: E0302 14:27:50.679195 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.679337 kubelet[2823]: W0302 14:27:50.679317 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.679424 kubelet[2823]: E0302 14:27:50.679406 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.688252 kubelet[2823]: E0302 14:27:50.687886 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.688252 kubelet[2823]: W0302 14:27:50.687913 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.688252 kubelet[2823]: E0302 14:27:50.688066 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.695389 kubelet[2823]: E0302 14:27:50.694671 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.695389 kubelet[2823]: W0302 14:27:50.694690 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.695389 kubelet[2823]: E0302 14:27:50.694709 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.695389 kubelet[2823]: E0302 14:27:50.695149 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.695389 kubelet[2823]: W0302 14:27:50.695164 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.695389 kubelet[2823]: E0302 14:27:50.695180 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.695389 kubelet[2823]: I0302 14:27:50.695332 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/f81f0355-d756-4ec2-b9d0-9639b176b2ea-varrun\") pod \"csi-node-driver-fgxfk\" (UID: \"f81f0355-d756-4ec2-b9d0-9639b176b2ea\") " pod="calico-system/csi-node-driver-fgxfk" Mar 2 14:27:50.702795 kubelet[2823]: E0302 14:27:50.702394 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.702795 kubelet[2823]: W0302 14:27:50.702424 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.702795 kubelet[2823]: E0302 14:27:50.702506 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.708821 kubelet[2823]: E0302 14:27:50.708800 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.709636 kubelet[2823]: W0302 14:27:50.709371 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.709636 kubelet[2823]: E0302 14:27:50.709400 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.709973 kubelet[2823]: E0302 14:27:50.709916 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.710638 kubelet[2823]: W0302 14:27:50.710496 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.710638 kubelet[2823]: E0302 14:27:50.710523 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.711707 kubelet[2823]: E0302 14:27:50.711691 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.711792 kubelet[2823]: W0302 14:27:50.711778 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.725736 kubelet[2823]: E0302 14:27:50.725650 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.729066 kubelet[2823]: E0302 14:27:50.728965 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.729382 kubelet[2823]: W0302 14:27:50.729290 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.729661 kubelet[2823]: E0302 14:27:50.729555 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.730236 kubelet[2823]: E0302 14:27:50.730219 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.730336 kubelet[2823]: W0302 14:27:50.730318 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.730429 kubelet[2823]: E0302 14:27:50.730414 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.730971 kubelet[2823]: E0302 14:27:50.730917 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.731064 kubelet[2823]: W0302 14:27:50.731049 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.731141 kubelet[2823]: E0302 14:27:50.731124 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.732983 kubelet[2823]: E0302 14:27:50.732964 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.734023 kubelet[2823]: W0302 14:27:50.733973 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.734023 kubelet[2823]: E0302 14:27:50.734000 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.735617 kubelet[2823]: I0302 14:27:50.735593 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9ctw\" (UniqueName: \"kubernetes.io/projected/f81f0355-d756-4ec2-b9d0-9639b176b2ea-kube-api-access-g9ctw\") pod \"csi-node-driver-fgxfk\" (UID: \"f81f0355-d756-4ec2-b9d0-9639b176b2ea\") " pod="calico-system/csi-node-driver-fgxfk" Mar 2 14:27:50.737426 systemd[1]: Started cri-containerd-5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b.scope - libcontainer container 5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b. Mar 2 14:27:50.742594 kubelet[2823]: E0302 14:27:50.742570 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.743849 kubelet[2823]: W0302 14:27:50.743730 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.746153 kubelet[2823]: E0302 14:27:50.744241 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.750885 kubelet[2823]: E0302 14:27:50.750731 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.753747 kubelet[2823]: W0302 14:27:50.752764 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.753747 kubelet[2823]: E0302 14:27:50.752796 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.757007 kubelet[2823]: E0302 14:27:50.756622 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.757007 kubelet[2823]: W0302 14:27:50.756644 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.757007 kubelet[2823]: E0302 14:27:50.756662 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.765140 kubelet[2823]: E0302 14:27:50.764738 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.765140 kubelet[2823]: W0302 14:27:50.764762 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.765140 kubelet[2823]: E0302 14:27:50.764785 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.765685 kubelet[2823]: E0302 14:27:50.765666 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.768202 kubelet[2823]: W0302 14:27:50.768131 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.771019 kubelet[2823]: E0302 14:27:50.770963 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.783521 kubelet[2823]: E0302 14:27:50.783347 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.783660 kubelet[2823]: W0302 14:27:50.783641 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.783729 kubelet[2823]: E0302 14:27:50.783716 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.796386 kubelet[2823]: E0302 14:27:50.796358 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.797108 kubelet[2823]: W0302 14:27:50.796611 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.797108 kubelet[2823]: E0302 14:27:50.796640 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.802038 kubelet[2823]: E0302 14:27:50.802016 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.802295 kubelet[2823]: W0302 14:27:50.802150 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.802295 kubelet[2823]: E0302 14:27:50.802180 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.809182 kubelet[2823]: E0302 14:27:50.809158 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.809512 kubelet[2823]: W0302 14:27:50.809302 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.809512 kubelet[2823]: E0302 14:27:50.809329 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.810167 kubelet[2823]: E0302 14:27:50.810149 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.810255 kubelet[2823]: W0302 14:27:50.810240 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.810323 kubelet[2823]: E0302 14:27:50.810309 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.832608 kubelet[2823]: E0302 14:27:50.830113 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.832608 kubelet[2823]: W0302 14:27:50.832323 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.832608 kubelet[2823]: E0302 14:27:50.832373 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.836658 kubelet[2823]: E0302 14:27:50.836580 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.836658 kubelet[2823]: W0302 14:27:50.836604 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.836658 kubelet[2823]: E0302 14:27:50.836630 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.838894 kubelet[2823]: E0302 14:27:50.838875 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.840649 kubelet[2823]: W0302 14:27:50.840279 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.840649 kubelet[2823]: E0302 14:27:50.840309 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.846132 kubelet[2823]: E0302 14:27:50.846097 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.846132 kubelet[2823]: W0302 14:27:50.846329 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.846132 kubelet[2823]: E0302 14:27:50.846362 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.852303 kubelet[2823]: E0302 14:27:50.850544 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.852303 kubelet[2823]: W0302 14:27:50.850579 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.852303 kubelet[2823]: E0302 14:27:50.850608 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.854362 kubelet[2823]: E0302 14:27:50.854342 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.854546 kubelet[2823]: W0302 14:27:50.854526 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.854639 kubelet[2823]: E0302 14:27:50.854622 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.855167 kubelet[2823]: E0302 14:27:50.855146 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.855327 kubelet[2823]: W0302 14:27:50.855240 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.855327 kubelet[2823]: E0302 14:27:50.855262 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.855860 kubelet[2823]: E0302 14:27:50.855770 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.855860 kubelet[2823]: W0302 14:27:50.855789 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.855860 kubelet[2823]: E0302 14:27:50.855804 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.858626 kubelet[2823]: E0302 14:27:50.858428 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.859281 kubelet[2823]: W0302 14:27:50.858814 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.859281 kubelet[2823]: E0302 14:27:50.858836 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.862848 kubelet[2823]: E0302 14:27:50.862827 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.865152 kubelet[2823]: W0302 14:27:50.865124 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.867703 kubelet[2823]: E0302 14:27:50.867295 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.873081 kubelet[2823]: E0302 14:27:50.872864 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.873969 kubelet[2823]: W0302 14:27:50.873499 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.873969 kubelet[2823]: E0302 14:27:50.873533 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.879504 kubelet[2823]: E0302 14:27:50.877332 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.879504 kubelet[2823]: W0302 14:27:50.877630 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.879504 kubelet[2823]: E0302 14:27:50.877654 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.941547 kubelet[2823]: E0302 14:27:50.941199 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.941547 kubelet[2823]: W0302 14:27:50.941228 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.941547 kubelet[2823]: E0302 14:27:50.941251 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.942493 kubelet[2823]: E0302 14:27:50.942059 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.942493 kubelet[2823]: W0302 14:27:50.942078 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.942493 kubelet[2823]: E0302 14:27:50.942093 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.947741 kubelet[2823]: E0302 14:27:50.947082 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.947741 kubelet[2823]: W0302 14:27:50.947101 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.947741 kubelet[2823]: E0302 14:27:50.947193 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.951872 kubelet[2823]: E0302 14:27:50.951853 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.955714 kubelet[2823]: W0302 14:27:50.955514 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.955714 kubelet[2823]: E0302 14:27:50.955543 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.958421 kubelet[2823]: E0302 14:27:50.955985 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.958421 kubelet[2823]: W0302 14:27:50.956002 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.958421 kubelet[2823]: E0302 14:27:50.956015 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.958990 kubelet[2823]: E0302 14:27:50.958864 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.959165 kubelet[2823]: W0302 14:27:50.959147 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.959416 kubelet[2823]: E0302 14:27:50.959399 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:50.961002 kubelet[2823]: E0302 14:27:50.960833 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:50.961002 kubelet[2823]: W0302 14:27:50.960852 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:50.961910 kubelet[2823]: E0302 14:27:50.960865 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:51.000730 kubelet[2823]: E0302 14:27:50.997687 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:51.000730 kubelet[2823]: W0302 14:27:50.997807 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:51.000730 kubelet[2823]: E0302 14:27:50.997916 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:51.001063 kubelet[2823]: E0302 14:27:51.000859 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:51.001063 kubelet[2823]: W0302 14:27:51.000884 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:51.005785 kubelet[2823]: E0302 14:27:51.000911 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:51.007893 kubelet[2823]: E0302 14:27:51.007672 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:51.007893 kubelet[2823]: W0302 14:27:51.007698 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:51.007893 kubelet[2823]: E0302 14:27:51.007722 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:51.133728 containerd[1574]: time="2026-03-02T14:27:51.128585929Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-mbhdq,Uid:89a4e8cd-0f0e-4f2b-912c-3aceb3d7661a,Namespace:calico-system,Attempt:0,} returns sandbox id \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\"" Mar 2 14:27:51.150344 kubelet[2823]: E0302 14:27:51.130400 2823 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 14:27:51.150344 kubelet[2823]: W0302 14:27:51.130423 2823 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 14:27:51.150344 kubelet[2823]: E0302 14:27:51.130514 2823 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 14:27:51.167518 containerd[1574]: time="2026-03-02T14:27:51.161356043Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\"" Mar 2 14:27:51.328018 containerd[1574]: time="2026-03-02T14:27:51.322190379Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-78dff98b8f-fh7br,Uid:404f3cd9-5a00-48bc-b7c9-ef7413827faf,Namespace:calico-system,Attempt:0,} returns sandbox id \"c25234fbbf3c026a1878d06769dbd0e187e30eb05bca8798541087348e8b5c45\"" Mar 2 14:27:51.332078 kubelet[2823]: E0302 14:27:51.330909 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:27:52.101756 kubelet[2823]: E0302 14:27:52.101687 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:27:54.098515 kubelet[2823]: E0302 14:27:54.098342 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:27:54.698715 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount124744596.mount: Deactivated successfully. Mar 2 14:27:56.111495 kubelet[2823]: E0302 14:27:56.099000 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:27:57.937917 containerd[1574]: time="2026-03-02T14:27:57.937692118Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:27:57.958128 containerd[1574]: time="2026-03-02T14:27:57.946922012Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3: active requests=0, bytes read=6186335" Mar 2 14:27:57.967766 containerd[1574]: time="2026-03-02T14:27:57.967154892Z" level=info msg="ImageCreate event name:\"sha256:ecc2a8ca795d595c3a806abf201d701228ddc7a8373e906441c9470dfeadd022\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:27:58.105231 kubelet[2823]: E0302 14:27:58.104239 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:27:58.107352 containerd[1574]: time="2026-03-02T14:27:58.107101697Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\" with image id \"sha256:ecc2a8ca795d595c3a806abf201d701228ddc7a8373e906441c9470dfeadd022\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:6cdc6cc2f7cdcbd4bf2d9b6a59c03ed98b5c47f22e467d78b5c06e5fd7bff132\", size \"6186157\" in 6.945698876s" Mar 2 14:27:58.107352 containerd[1574]: time="2026-03-02T14:27:58.107175774Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\" returns image reference \"sha256:ecc2a8ca795d595c3a806abf201d701228ddc7a8373e906441c9470dfeadd022\"" Mar 2 14:27:58.115276 containerd[1574]: time="2026-03-02T14:27:58.114826393Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:6cdc6cc2f7cdcbd4bf2d9b6a59c03ed98b5c47f22e467d78b5c06e5fd7bff132\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:27:58.303576 containerd[1574]: time="2026-03-02T14:27:58.303328687Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.3\"" Mar 2 14:27:58.380757 containerd[1574]: time="2026-03-02T14:27:58.380494446Z" level=info msg="CreateContainer within sandbox \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 2 14:27:58.820250 containerd[1574]: time="2026-03-02T14:27:58.820049844Z" level=info msg="Container ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:27:58.944071 containerd[1574]: time="2026-03-02T14:27:58.943828848Z" level=info msg="CreateContainer within sandbox \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103\"" Mar 2 14:27:58.946925 containerd[1574]: time="2026-03-02T14:27:58.945976516Z" level=info msg="StartContainer for \"ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103\"" Mar 2 14:27:58.966496 containerd[1574]: time="2026-03-02T14:27:58.965761139Z" level=info msg="connecting to shim ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103" address="unix:///run/containerd/s/0b9e2f1b1317597f174e05833aa8896ff5dba404179e70ab1652b5f1b44c0e65" protocol=ttrpc version=3 Mar 2 14:27:59.475799 systemd[1]: Started cri-containerd-ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103.scope - libcontainer container ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103. Mar 2 14:27:59.925637 containerd[1574]: time="2026-03-02T14:27:59.925584301Z" level=info msg="StartContainer for \"ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103\" returns successfully" Mar 2 14:28:00.010897 systemd[1]: cri-containerd-ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103.scope: Deactivated successfully. Mar 2 14:28:00.033647 containerd[1574]: time="2026-03-02T14:28:00.033536648Z" level=info msg="received container exit event container_id:\"ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103\" id:\"ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103\" pid:3602 exited_at:{seconds:1772461680 nanos:19612157}" Mar 2 14:28:00.099788 kubelet[2823]: E0302 14:28:00.098951 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:00.216840 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ae228baa3e566ce83fcf75bcea5a3c507c47cbe5f6347750fc0b8a08193bb103-rootfs.mount: Deactivated successfully. Mar 2 14:28:02.102793 kubelet[2823]: E0302 14:28:02.098357 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:04.103925 kubelet[2823]: E0302 14:28:04.100107 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:06.121228 kubelet[2823]: E0302 14:28:06.120722 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:07.057308 containerd[1574]: time="2026-03-02T14:28:07.057106111Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:07.063361 containerd[1574]: time="2026-03-02T14:28:07.060970203Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.3: active requests=0, bytes read=34538513" Mar 2 14:28:07.064326 containerd[1574]: time="2026-03-02T14:28:07.064250520Z" level=info msg="ImageCreate event name:\"sha256:0aa5de4a226c8dff91be273305b5e55a8b7019ef516599fd15c7e4434085cd65\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:07.074991 containerd[1574]: time="2026-03-02T14:28:07.073143052Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:3e62cf98a20c42a1786397d0192cfb639634ef95c6f463ab92f0439a5c1a4ae5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:07.074991 containerd[1574]: time="2026-03-02T14:28:07.074363331Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.3\" with image id \"sha256:0aa5de4a226c8dff91be273305b5e55a8b7019ef516599fd15c7e4434085cd65\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:3e62cf98a20c42a1786397d0192cfb639634ef95c6f463ab92f0439a5c1a4ae5\", size \"36094550\" in 8.770179957s" Mar 2 14:28:07.074991 containerd[1574]: time="2026-03-02T14:28:07.074422882Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.3\" returns image reference \"sha256:0aa5de4a226c8dff91be273305b5e55a8b7019ef516599fd15c7e4434085cd65\"" Mar 2 14:28:07.088763 containerd[1574]: time="2026-03-02T14:28:07.088542336Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.3\"" Mar 2 14:28:07.194060 containerd[1574]: time="2026-03-02T14:28:07.194015243Z" level=info msg="CreateContainer within sandbox \"c25234fbbf3c026a1878d06769dbd0e187e30eb05bca8798541087348e8b5c45\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Mar 2 14:28:07.258959 containerd[1574]: time="2026-03-02T14:28:07.253633148Z" level=info msg="Container ee58d21bdb50ba10b667c431e688f5b1252c34666f288e334c25c7e77e62448d: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:28:07.314184 containerd[1574]: time="2026-03-02T14:28:07.314050776Z" level=info msg="CreateContainer within sandbox \"c25234fbbf3c026a1878d06769dbd0e187e30eb05bca8798541087348e8b5c45\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"ee58d21bdb50ba10b667c431e688f5b1252c34666f288e334c25c7e77e62448d\"" Mar 2 14:28:07.318944 containerd[1574]: time="2026-03-02T14:28:07.318352812Z" level=info msg="StartContainer for \"ee58d21bdb50ba10b667c431e688f5b1252c34666f288e334c25c7e77e62448d\"" Mar 2 14:28:07.332704 containerd[1574]: time="2026-03-02T14:28:07.332577483Z" level=info msg="connecting to shim ee58d21bdb50ba10b667c431e688f5b1252c34666f288e334c25c7e77e62448d" address="unix:///run/containerd/s/cd7c9448a7c15dbb4e19d27d729d5020e87b8f38c3d1ba02e9734cf48bb19525" protocol=ttrpc version=3 Mar 2 14:28:07.469123 systemd[1]: Started cri-containerd-ee58d21bdb50ba10b667c431e688f5b1252c34666f288e334c25c7e77e62448d.scope - libcontainer container ee58d21bdb50ba10b667c431e688f5b1252c34666f288e334c25c7e77e62448d. Mar 2 14:28:07.806868 containerd[1574]: time="2026-03-02T14:28:07.806694174Z" level=info msg="StartContainer for \"ee58d21bdb50ba10b667c431e688f5b1252c34666f288e334c25c7e77e62448d\" returns successfully" Mar 2 14:28:08.105008 kubelet[2823]: E0302 14:28:08.100593 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:08.730109 kubelet[2823]: E0302 14:28:08.729560 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:08.861175 kubelet[2823]: I0302 14:28:08.857651 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-78dff98b8f-fh7br" podStartSLOduration=7.107779292 podStartE2EDuration="22.85763086s" podCreationTimestamp="2026-03-02 14:27:46 +0000 UTC" firstStartedPulling="2026-03-02 14:27:51.332635573 +0000 UTC m=+34.766303576" lastFinishedPulling="2026-03-02 14:28:07.082487141 +0000 UTC m=+50.516155144" observedRunningTime="2026-03-02 14:28:08.857549369 +0000 UTC m=+52.291217392" watchObservedRunningTime="2026-03-02 14:28:08.85763086 +0000 UTC m=+52.291298863" Mar 2 14:28:09.746555 kubelet[2823]: E0302 14:28:09.746141 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:10.107689 kubelet[2823]: E0302 14:28:10.098088 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:10.761372 kubelet[2823]: E0302 14:28:10.759377 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:12.102172 kubelet[2823]: E0302 14:28:12.098968 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:14.103018 kubelet[2823]: E0302 14:28:14.102723 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:16.101243 kubelet[2823]: E0302 14:28:16.098614 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:18.115520 kubelet[2823]: E0302 14:28:18.115396 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:20.100043 kubelet[2823]: E0302 14:28:20.099638 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:22.098936 kubelet[2823]: E0302 14:28:22.098346 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:24.101685 kubelet[2823]: E0302 14:28:24.101200 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:26.111035 kubelet[2823]: E0302 14:28:26.105397 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:26.111581 kubelet[2823]: E0302 14:28:26.111554 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:28.099061 kubelet[2823]: E0302 14:28:28.098116 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:30.099318 kubelet[2823]: E0302 14:28:30.098160 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:32.099863 kubelet[2823]: E0302 14:28:32.099714 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:33.322758 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3389556817.mount: Deactivated successfully. Mar 2 14:28:33.442114 containerd[1574]: time="2026-03-02T14:28:33.441074098Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:33.449306 containerd[1574]: time="2026-03-02T14:28:33.448518064Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.3: active requests=0, bytes read=159483365" Mar 2 14:28:33.451743 containerd[1574]: time="2026-03-02T14:28:33.451647171Z" level=info msg="ImageCreate event name:\"sha256:f8495fa3f644ae70c7e5131c7baf23f80864678694dbf1a6a4d0557528433740\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:33.462679 containerd[1574]: time="2026-03-02T14:28:33.462567229Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:c7aefc80042b94800407ab45640b59402d2897ae8755b9d8370516e7b0e404bc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:33.538224 containerd[1574]: time="2026-03-02T14:28:33.538090857Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.3\" with image id \"sha256:f8495fa3f644ae70c7e5131c7baf23f80864678694dbf1a6a4d0557528433740\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:c7aefc80042b94800407ab45640b59402d2897ae8755b9d8370516e7b0e404bc\", size \"159483227\" in 26.449130903s" Mar 2 14:28:33.538224 containerd[1574]: time="2026-03-02T14:28:33.538173309Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.3\" returns image reference \"sha256:f8495fa3f644ae70c7e5131c7baf23f80864678694dbf1a6a4d0557528433740\"" Mar 2 14:28:33.586421 containerd[1574]: time="2026-03-02T14:28:33.582105607Z" level=info msg="CreateContainer within sandbox \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Mar 2 14:28:33.700581 containerd[1574]: time="2026-03-02T14:28:33.699983999Z" level=info msg="Container 461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:28:33.807738 containerd[1574]: time="2026-03-02T14:28:33.807628624Z" level=info msg="CreateContainer within sandbox \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f\"" Mar 2 14:28:33.808376 containerd[1574]: time="2026-03-02T14:28:33.808269527Z" level=info msg="StartContainer for \"461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f\"" Mar 2 14:28:33.811381 containerd[1574]: time="2026-03-02T14:28:33.811183987Z" level=info msg="connecting to shim 461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f" address="unix:///run/containerd/s/0b9e2f1b1317597f174e05833aa8896ff5dba404179e70ab1652b5f1b44c0e65" protocol=ttrpc version=3 Mar 2 14:28:33.907670 systemd[1]: Started cri-containerd-461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f.scope - libcontainer container 461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f. Mar 2 14:28:34.099693 kubelet[2823]: E0302 14:28:34.098591 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:34.266533 containerd[1574]: time="2026-03-02T14:28:34.265529778Z" level=info msg="StartContainer for \"461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f\" returns successfully" Mar 2 14:28:34.454430 systemd[1]: cri-containerd-461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f.scope: Deactivated successfully. Mar 2 14:28:34.465099 containerd[1574]: time="2026-03-02T14:28:34.461350151Z" level=info msg="received container exit event container_id:\"461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f\" id:\"461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f\" pid:3714 exited_at:{seconds:1772461714 nanos:460699152}" Mar 2 14:28:34.634079 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-461e61580afc1a2604ce01c23b9af840f8a234559cfba793a24752e1db87bf1f-rootfs.mount: Deactivated successfully. Mar 2 14:28:36.114172 kubelet[2823]: E0302 14:28:36.108942 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:36.165228 containerd[1574]: time="2026-03-02T14:28:36.165176035Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.3\"" Mar 2 14:28:38.101983 kubelet[2823]: E0302 14:28:38.101613 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:40.115626 kubelet[2823]: E0302 14:28:40.114494 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:41.099629 kubelet[2823]: E0302 14:28:41.099581 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:42.103909 kubelet[2823]: E0302 14:28:42.103851 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:44.098072 kubelet[2823]: E0302 14:28:44.098023 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:45.111215 kubelet[2823]: E0302 14:28:45.107766 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:46.098699 kubelet[2823]: E0302 14:28:46.098279 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:47.830505 containerd[1574]: time="2026-03-02T14:28:47.828921842Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:47.830505 containerd[1574]: time="2026-03-02T14:28:47.830463884Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.3: active requests=0, bytes read=70584418" Mar 2 14:28:47.834292 containerd[1574]: time="2026-03-02T14:28:47.834214352Z" level=info msg="ImageCreate event name:\"sha256:f2520fbaa2761d3cc6c294dcad9c4dc33442ee0c856af33cefd0da5346519691\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:47.847583 containerd[1574]: time="2026-03-02T14:28:47.847525525Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:c25deb6a4b79f5e595eb464adf9fb3735ea5623889e249d5b3efa0b42ffcbb47\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:28:47.853177 containerd[1574]: time="2026-03-02T14:28:47.852978627Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.3\" with image id \"sha256:f2520fbaa2761d3cc6c294dcad9c4dc33442ee0c856af33cefd0da5346519691\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:c25deb6a4b79f5e595eb464adf9fb3735ea5623889e249d5b3efa0b42ffcbb47\", size \"72140463\" in 11.687006028s" Mar 2 14:28:47.853177 containerd[1574]: time="2026-03-02T14:28:47.853045892Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.3\" returns image reference \"sha256:f2520fbaa2761d3cc6c294dcad9c4dc33442ee0c856af33cefd0da5346519691\"" Mar 2 14:28:47.881898 containerd[1574]: time="2026-03-02T14:28:47.881754769Z" level=info msg="CreateContainer within sandbox \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 2 14:28:47.986289 containerd[1574]: time="2026-03-02T14:28:47.982739147Z" level=info msg="Container fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:28:48.031899 containerd[1574]: time="2026-03-02T14:28:48.031081154Z" level=info msg="CreateContainer within sandbox \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375\"" Mar 2 14:28:48.034241 containerd[1574]: time="2026-03-02T14:28:48.033365931Z" level=info msg="StartContainer for \"fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375\"" Mar 2 14:28:48.039938 containerd[1574]: time="2026-03-02T14:28:48.037706088Z" level=info msg="connecting to shim fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375" address="unix:///run/containerd/s/0b9e2f1b1317597f174e05833aa8896ff5dba404179e70ab1652b5f1b44c0e65" protocol=ttrpc version=3 Mar 2 14:28:48.098904 kubelet[2823]: E0302 14:28:48.098376 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:48.153730 systemd[1]: Started cri-containerd-fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375.scope - libcontainer container fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375. Mar 2 14:28:48.723037 containerd[1574]: time="2026-03-02T14:28:48.721379644Z" level=info msg="StartContainer for \"fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375\" returns successfully" Mar 2 14:28:49.104996 kubelet[2823]: E0302 14:28:49.103847 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:50.108095 kubelet[2823]: E0302 14:28:50.105096 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:52.066694 systemd[1]: cri-containerd-fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375.scope: Deactivated successfully. Mar 2 14:28:52.068881 containerd[1574]: time="2026-03-02T14:28:52.068712564Z" level=info msg="received container exit event container_id:\"fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375\" id:\"fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375\" pid:3771 exited_at:{seconds:1772461732 nanos:68373522}" Mar 2 14:28:52.083271 systemd[1]: cri-containerd-fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375.scope: Consumed 1.455s CPU time, 187.1M memory peak, 5.9M read from disk, 176.9M written to disk. Mar 2 14:28:52.098953 kubelet[2823]: E0302 14:28:52.098530 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:52.283397 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fff06c6c03de61a553ef8843575de04ce285c28a4fab0e613173ea0b090ee375-rootfs.mount: Deactivated successfully. Mar 2 14:28:52.310924 kubelet[2823]: I0302 14:28:52.309318 2823 kubelet_node_status.go:439] "Fast updating node status as it just became ready" Mar 2 14:28:52.648559 containerd[1574]: time="2026-03-02T14:28:52.647871407Z" level=info msg="CreateContainer within sandbox \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 2 14:28:52.733714 systemd[1]: Created slice kubepods-besteffort-pod7633016e_887b_423d_9baf_5d37aa0a7844.slice - libcontainer container kubepods-besteffort-pod7633016e_887b_423d_9baf_5d37aa0a7844.slice. Mar 2 14:28:52.765989 containerd[1574]: time="2026-03-02T14:28:52.764757320Z" level=info msg="Container e0d42c8163228963da3623906bc9e4658c9f8028164fab775fd3a6db322109f9: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:28:52.806958 systemd[1]: Created slice kubepods-besteffort-pod0dd516cb_8822_4937_b80a_6a7395992c08.slice - libcontainer container kubepods-besteffort-pod0dd516cb_8822_4937_b80a_6a7395992c08.slice. Mar 2 14:28:52.821834 kubelet[2823]: I0302 14:28:52.821029 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7633016e-887b-423d-9baf-5d37aa0a7844-tigera-ca-bundle\") pod \"calico-kube-controllers-6d7c47ccd9-vmfz4\" (UID: \"7633016e-887b-423d-9baf-5d37aa0a7844\") " pod="calico-system/calico-kube-controllers-6d7c47ccd9-vmfz4" Mar 2 14:28:52.821834 kubelet[2823]: I0302 14:28:52.821198 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhcl7\" (UniqueName: \"kubernetes.io/projected/7633016e-887b-423d-9baf-5d37aa0a7844-kube-api-access-xhcl7\") pod \"calico-kube-controllers-6d7c47ccd9-vmfz4\" (UID: \"7633016e-887b-423d-9baf-5d37aa0a7844\") " pod="calico-system/calico-kube-controllers-6d7c47ccd9-vmfz4" Mar 2 14:28:52.853237 systemd[1]: Created slice kubepods-burstable-pod71148f0a_7223_4509_99b6_37365349cb1b.slice - libcontainer container kubepods-burstable-pod71148f0a_7223_4509_99b6_37365349cb1b.slice. Mar 2 14:28:52.878518 containerd[1574]: time="2026-03-02T14:28:52.876013376Z" level=info msg="CreateContainer within sandbox \"5cec42d35340803efba46a633276b56fa8a1299c9f6ff26308a588e2bc194c3b\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"e0d42c8163228963da3623906bc9e4658c9f8028164fab775fd3a6db322109f9\"" Mar 2 14:28:52.880417 containerd[1574]: time="2026-03-02T14:28:52.880145141Z" level=info msg="StartContainer for \"e0d42c8163228963da3623906bc9e4658c9f8028164fab775fd3a6db322109f9\"" Mar 2 14:28:52.888574 systemd[1]: Created slice kubepods-besteffort-pod1acbed91_a452_4e11_a018_a8670f52ea1a.slice - libcontainer container kubepods-besteffort-pod1acbed91_a452_4e11_a018_a8670f52ea1a.slice. Mar 2 14:28:52.912045 containerd[1574]: time="2026-03-02T14:28:52.904871076Z" level=info msg="connecting to shim e0d42c8163228963da3623906bc9e4658c9f8028164fab775fd3a6db322109f9" address="unix:///run/containerd/s/0b9e2f1b1317597f174e05833aa8896ff5dba404179e70ab1652b5f1b44c0e65" protocol=ttrpc version=3 Mar 2 14:28:52.958067 systemd[1]: Created slice kubepods-burstable-podbc8db32a_755c_44fc_abe9_e927df38b7b5.slice - libcontainer container kubepods-burstable-podbc8db32a_755c_44fc_abe9_e927df38b7b5.slice. Mar 2 14:28:52.964872 kubelet[2823]: I0302 14:28:52.956753 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/a5277f53-71c0-4d52-bac9-fd4719f10daa-goldmane-key-pair\") pod \"goldmane-54d7f6b6d6-zx66p\" (UID: \"a5277f53-71c0-4d52-bac9-fd4719f10daa\") " pod="calico-system/goldmane-54d7f6b6d6-zx66p" Mar 2 14:28:52.964872 kubelet[2823]: I0302 14:28:52.961177 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1acbed91-a452-4e11-a018-a8670f52ea1a-whisker-ca-bundle\") pod \"whisker-686f4f4697-mjrdf\" (UID: \"1acbed91-a452-4e11-a018-a8670f52ea1a\") " pod="calico-system/whisker-686f4f4697-mjrdf" Mar 2 14:28:52.964872 kubelet[2823]: I0302 14:28:52.962509 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvpfh\" (UniqueName: \"kubernetes.io/projected/1acbed91-a452-4e11-a018-a8670f52ea1a-kube-api-access-bvpfh\") pod \"whisker-686f4f4697-mjrdf\" (UID: \"1acbed91-a452-4e11-a018-a8670f52ea1a\") " pod="calico-system/whisker-686f4f4697-mjrdf" Mar 2 14:28:52.965366 kubelet[2823]: I0302 14:28:52.965312 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxgv6\" (UniqueName: \"kubernetes.io/projected/a5277f53-71c0-4d52-bac9-fd4719f10daa-kube-api-access-pxgv6\") pod \"goldmane-54d7f6b6d6-zx66p\" (UID: \"a5277f53-71c0-4d52-bac9-fd4719f10daa\") " pod="calico-system/goldmane-54d7f6b6d6-zx66p" Mar 2 14:28:52.965623 kubelet[2823]: I0302 14:28:52.965424 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/1acbed91-a452-4e11-a018-a8670f52ea1a-nginx-config\") pod \"whisker-686f4f4697-mjrdf\" (UID: \"1acbed91-a452-4e11-a018-a8670f52ea1a\") " pod="calico-system/whisker-686f4f4697-mjrdf" Mar 2 14:28:52.965623 kubelet[2823]: I0302 14:28:52.965572 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/1acbed91-a452-4e11-a018-a8670f52ea1a-whisker-backend-key-pair\") pod \"whisker-686f4f4697-mjrdf\" (UID: \"1acbed91-a452-4e11-a018-a8670f52ea1a\") " pod="calico-system/whisker-686f4f4697-mjrdf" Mar 2 14:28:52.965709 kubelet[2823]: I0302 14:28:52.965677 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71148f0a-7223-4509-99b6-37365349cb1b-config-volume\") pod \"coredns-66bc5c9577-xrdgw\" (UID: \"71148f0a-7223-4509-99b6-37365349cb1b\") " pod="kube-system/coredns-66bc5c9577-xrdgw" Mar 2 14:28:52.965789 kubelet[2823]: I0302 14:28:52.965705 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkk89\" (UniqueName: \"kubernetes.io/projected/71148f0a-7223-4509-99b6-37365349cb1b-kube-api-access-zkk89\") pod \"coredns-66bc5c9577-xrdgw\" (UID: \"71148f0a-7223-4509-99b6-37365349cb1b\") " pod="kube-system/coredns-66bc5c9577-xrdgw" Mar 2 14:28:52.965897 kubelet[2823]: I0302 14:28:52.965844 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc8db32a-755c-44fc-abe9-e927df38b7b5-config-volume\") pod \"coredns-66bc5c9577-xlqmd\" (UID: \"bc8db32a-755c-44fc-abe9-e927df38b7b5\") " pod="kube-system/coredns-66bc5c9577-xlqmd" Mar 2 14:28:52.966053 kubelet[2823]: I0302 14:28:52.965974 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/ceacecf8-f59e-4263-8232-2bc003edccd4-calico-apiserver-certs\") pod \"calico-apiserver-68647fc8d9-p874s\" (UID: \"ceacecf8-f59e-4263-8232-2bc003edccd4\") " pod="calico-system/calico-apiserver-68647fc8d9-p874s" Mar 2 14:28:52.966230 kubelet[2823]: I0302 14:28:52.966173 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5h75\" (UniqueName: \"kubernetes.io/projected/bc8db32a-755c-44fc-abe9-e927df38b7b5-kube-api-access-v5h75\") pod \"coredns-66bc5c9577-xlqmd\" (UID: \"bc8db32a-755c-44fc-abe9-e927df38b7b5\") " pod="kube-system/coredns-66bc5c9577-xlqmd" Mar 2 14:28:52.966402 kubelet[2823]: I0302 14:28:52.966345 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/0dd516cb-8822-4937-b80a-6a7395992c08-calico-apiserver-certs\") pod \"calico-apiserver-68647fc8d9-mvfxc\" (UID: \"0dd516cb-8822-4937-b80a-6a7395992c08\") " pod="calico-system/calico-apiserver-68647fc8d9-mvfxc" Mar 2 14:28:52.966627 kubelet[2823]: I0302 14:28:52.966586 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frb7v\" (UniqueName: \"kubernetes.io/projected/ceacecf8-f59e-4263-8232-2bc003edccd4-kube-api-access-frb7v\") pod \"calico-apiserver-68647fc8d9-p874s\" (UID: \"ceacecf8-f59e-4263-8232-2bc003edccd4\") " pod="calico-system/calico-apiserver-68647fc8d9-p874s" Mar 2 14:28:52.966769 kubelet[2823]: I0302 14:28:52.966629 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdmvm\" (UniqueName: \"kubernetes.io/projected/0dd516cb-8822-4937-b80a-6a7395992c08-kube-api-access-qdmvm\") pod \"calico-apiserver-68647fc8d9-mvfxc\" (UID: \"0dd516cb-8822-4937-b80a-6a7395992c08\") " pod="calico-system/calico-apiserver-68647fc8d9-mvfxc" Mar 2 14:28:52.966968 kubelet[2823]: I0302 14:28:52.966777 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5277f53-71c0-4d52-bac9-fd4719f10daa-config\") pod \"goldmane-54d7f6b6d6-zx66p\" (UID: \"a5277f53-71c0-4d52-bac9-fd4719f10daa\") " pod="calico-system/goldmane-54d7f6b6d6-zx66p" Mar 2 14:28:52.967040 kubelet[2823]: I0302 14:28:52.966986 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5277f53-71c0-4d52-bac9-fd4719f10daa-goldmane-ca-bundle\") pod \"goldmane-54d7f6b6d6-zx66p\" (UID: \"a5277f53-71c0-4d52-bac9-fd4719f10daa\") " pod="calico-system/goldmane-54d7f6b6d6-zx66p" Mar 2 14:28:53.002039 systemd[1]: Created slice kubepods-besteffort-poda5277f53_71c0_4d52_bac9_fd4719f10daa.slice - libcontainer container kubepods-besteffort-poda5277f53_71c0_4d52_bac9_fd4719f10daa.slice. Mar 2 14:28:53.023956 systemd[1]: Created slice kubepods-besteffort-podceacecf8_f59e_4263_8232_2bc003edccd4.slice - libcontainer container kubepods-besteffort-podceacecf8_f59e_4263_8232_2bc003edccd4.slice. Mar 2 14:28:53.054708 systemd[1]: Started cri-containerd-e0d42c8163228963da3623906bc9e4658c9f8028164fab775fd3a6db322109f9.scope - libcontainer container e0d42c8163228963da3623906bc9e4658c9f8028164fab775fd3a6db322109f9. Mar 2 14:28:53.316563 containerd[1574]: time="2026-03-02T14:28:53.316319371Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-686f4f4697-mjrdf,Uid:1acbed91-a452-4e11-a018-a8670f52ea1a,Namespace:calico-system,Attempt:0,}" Mar 2 14:28:53.319068 kubelet[2823]: E0302 14:28:53.318219 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:53.328119 containerd[1574]: time="2026-03-02T14:28:53.324649525Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-xlqmd,Uid:bc8db32a-755c-44fc-abe9-e927df38b7b5,Namespace:kube-system,Attempt:0,}" Mar 2 14:28:53.363499 containerd[1574]: time="2026-03-02T14:28:53.363184260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6d7c47ccd9-vmfz4,Uid:7633016e-887b-423d-9baf-5d37aa0a7844,Namespace:calico-system,Attempt:0,}" Mar 2 14:28:53.388713 containerd[1574]: time="2026-03-02T14:28:53.388638676Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68647fc8d9-p874s,Uid:ceacecf8-f59e-4263-8232-2bc003edccd4,Namespace:calico-system,Attempt:0,}" Mar 2 14:28:53.442514 containerd[1574]: time="2026-03-02T14:28:53.442402090Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68647fc8d9-mvfxc,Uid:0dd516cb-8822-4937-b80a-6a7395992c08,Namespace:calico-system,Attempt:0,}" Mar 2 14:28:53.442999 containerd[1574]: time="2026-03-02T14:28:53.442971661Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d7f6b6d6-zx66p,Uid:a5277f53-71c0-4d52-bac9-fd4719f10daa,Namespace:calico-system,Attempt:0,}" Mar 2 14:28:53.493162 kubelet[2823]: E0302 14:28:53.489129 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:28:53.499067 containerd[1574]: time="2026-03-02T14:28:53.498982353Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-xrdgw,Uid:71148f0a-7223-4509-99b6-37365349cb1b,Namespace:kube-system,Attempt:0,}" Mar 2 14:28:54.055240 containerd[1574]: time="2026-03-02T14:28:54.055062584Z" level=info msg="StartContainer for \"e0d42c8163228963da3623906bc9e4658c9f8028164fab775fd3a6db322109f9\" returns successfully" Mar 2 14:28:54.129006 systemd[1]: Created slice kubepods-besteffort-podf81f0355_d756_4ec2_b9d0_9639b176b2ea.slice - libcontainer container kubepods-besteffort-podf81f0355_d756_4ec2_b9d0_9639b176b2ea.slice. Mar 2 14:28:54.164626 containerd[1574]: time="2026-03-02T14:28:54.161523430Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fgxfk,Uid:f81f0355-d756-4ec2-b9d0-9639b176b2ea,Namespace:calico-system,Attempt:0,}" Mar 2 14:28:54.401835 containerd[1574]: time="2026-03-02T14:28:54.401651356Z" level=error msg="Failed to destroy network for sandbox \"7b3a79d252235e16ed7bcaa445688f5411a3497e7fb8d13cc1cae422e1366689\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.413697 systemd[1]: run-netns-cni\x2d004eba11\x2d637d\x2d4e7b\x2d7e96\x2d5e9f93e86d08.mount: Deactivated successfully. Mar 2 14:28:54.427348 containerd[1574]: time="2026-03-02T14:28:54.426791619Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-xlqmd,Uid:bc8db32a-755c-44fc-abe9-e927df38b7b5,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b3a79d252235e16ed7bcaa445688f5411a3497e7fb8d13cc1cae422e1366689\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.449764 containerd[1574]: time="2026-03-02T14:28:54.449704017Z" level=error msg="Failed to destroy network for sandbox \"4907142dc9f0c789b970929c448c7d6ac13a371394f024e5dd5a3baa5cb1a19a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.456228 kubelet[2823]: E0302 14:28:54.456020 2823 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b3a79d252235e16ed7bcaa445688f5411a3497e7fb8d13cc1cae422e1366689\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.456228 kubelet[2823]: E0302 14:28:54.456212 2823 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b3a79d252235e16ed7bcaa445688f5411a3497e7fb8d13cc1cae422e1366689\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-xlqmd" Mar 2 14:28:54.458765 kubelet[2823]: E0302 14:28:54.456244 2823 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b3a79d252235e16ed7bcaa445688f5411a3497e7fb8d13cc1cae422e1366689\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-xlqmd" Mar 2 14:28:54.458765 kubelet[2823]: E0302 14:28:54.456576 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-xlqmd_kube-system(bc8db32a-755c-44fc-abe9-e927df38b7b5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-xlqmd_kube-system(bc8db32a-755c-44fc-abe9-e927df38b7b5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7b3a79d252235e16ed7bcaa445688f5411a3497e7fb8d13cc1cae422e1366689\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-xlqmd" podUID="bc8db32a-755c-44fc-abe9-e927df38b7b5" Mar 2 14:28:54.468023 systemd[1]: run-netns-cni\x2d620a7ad9\x2d5cef\x2dd794\x2d2290\x2d1265da562fa5.mount: Deactivated successfully. Mar 2 14:28:54.492355 containerd[1574]: time="2026-03-02T14:28:54.491937913Z" level=error msg="Failed to destroy network for sandbox \"aab7e3d594d57f379a48a06849612c2ecc9cec990c079ad06ee4c8766c479b4b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.506548 systemd[1]: run-netns-cni\x2df157d517\x2d0ae5\x2d024d\x2d51bc\x2def5a5be515b6.mount: Deactivated successfully. Mar 2 14:28:54.525518 containerd[1574]: time="2026-03-02T14:28:54.525390537Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68647fc8d9-p874s,Uid:ceacecf8-f59e-4263-8232-2bc003edccd4,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4907142dc9f0c789b970929c448c7d6ac13a371394f024e5dd5a3baa5cb1a19a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.527497 containerd[1574]: time="2026-03-02T14:28:54.527387348Z" level=error msg="Failed to destroy network for sandbox \"fea5c6f9a46377c9c8d9fc0b555e5c39379c4e1abf8337bb918ff3624a1e2cf1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.537074 kubelet[2823]: E0302 14:28:54.530594 2823 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4907142dc9f0c789b970929c448c7d6ac13a371394f024e5dd5a3baa5cb1a19a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.537074 kubelet[2823]: E0302 14:28:54.530885 2823 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4907142dc9f0c789b970929c448c7d6ac13a371394f024e5dd5a3baa5cb1a19a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-68647fc8d9-p874s" Mar 2 14:28:54.537074 kubelet[2823]: E0302 14:28:54.530998 2823 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4907142dc9f0c789b970929c448c7d6ac13a371394f024e5dd5a3baa5cb1a19a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-68647fc8d9-p874s" Mar 2 14:28:54.537272 containerd[1574]: time="2026-03-02T14:28:54.533546577Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-686f4f4697-mjrdf,Uid:1acbed91-a452-4e11-a018-a8670f52ea1a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"aab7e3d594d57f379a48a06849612c2ecc9cec990c079ad06ee4c8766c479b4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.537387 kubelet[2823]: E0302 14:28:54.532253 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-68647fc8d9-p874s_calico-system(ceacecf8-f59e-4263-8232-2bc003edccd4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-68647fc8d9-p874s_calico-system(ceacecf8-f59e-4263-8232-2bc003edccd4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4907142dc9f0c789b970929c448c7d6ac13a371394f024e5dd5a3baa5cb1a19a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-68647fc8d9-p874s" podUID="ceacecf8-f59e-4263-8232-2bc003edccd4" Mar 2 14:28:54.541615 kubelet[2823]: E0302 14:28:54.538537 2823 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aab7e3d594d57f379a48a06849612c2ecc9cec990c079ad06ee4c8766c479b4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.541615 kubelet[2823]: E0302 14:28:54.540087 2823 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aab7e3d594d57f379a48a06849612c2ecc9cec990c079ad06ee4c8766c479b4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-686f4f4697-mjrdf" Mar 2 14:28:54.541722 kubelet[2823]: E0302 14:28:54.540698 2823 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aab7e3d594d57f379a48a06849612c2ecc9cec990c079ad06ee4c8766c479b4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-686f4f4697-mjrdf" Mar 2 14:28:54.543613 kubelet[2823]: E0302 14:28:54.542007 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-686f4f4697-mjrdf_calico-system(1acbed91-a452-4e11-a018-a8670f52ea1a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-686f4f4697-mjrdf_calico-system(1acbed91-a452-4e11-a018-a8670f52ea1a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"aab7e3d594d57f379a48a06849612c2ecc9cec990c079ad06ee4c8766c479b4b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-686f4f4697-mjrdf" podUID="1acbed91-a452-4e11-a018-a8670f52ea1a" Mar 2 14:28:54.544848 systemd[1]: run-netns-cni\x2d37f4bb0e\x2d13f0\x2df32b\x2d9716\x2db91b02eb81d3.mount: Deactivated successfully. Mar 2 14:28:54.563696 containerd[1574]: time="2026-03-02T14:28:54.563623874Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6d7c47ccd9-vmfz4,Uid:7633016e-887b-423d-9baf-5d37aa0a7844,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"fea5c6f9a46377c9c8d9fc0b555e5c39379c4e1abf8337bb918ff3624a1e2cf1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.566981 kubelet[2823]: E0302 14:28:54.566936 2823 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fea5c6f9a46377c9c8d9fc0b555e5c39379c4e1abf8337bb918ff3624a1e2cf1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.567145 kubelet[2823]: E0302 14:28:54.567121 2823 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fea5c6f9a46377c9c8d9fc0b555e5c39379c4e1abf8337bb918ff3624a1e2cf1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6d7c47ccd9-vmfz4" Mar 2 14:28:54.567253 kubelet[2823]: E0302 14:28:54.567220 2823 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fea5c6f9a46377c9c8d9fc0b555e5c39379c4e1abf8337bb918ff3624a1e2cf1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6d7c47ccd9-vmfz4" Mar 2 14:28:54.567410 kubelet[2823]: E0302 14:28:54.567369 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6d7c47ccd9-vmfz4_calico-system(7633016e-887b-423d-9baf-5d37aa0a7844)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6d7c47ccd9-vmfz4_calico-system(7633016e-887b-423d-9baf-5d37aa0a7844)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fea5c6f9a46377c9c8d9fc0b555e5c39379c4e1abf8337bb918ff3624a1e2cf1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6d7c47ccd9-vmfz4" podUID="7633016e-887b-423d-9baf-5d37aa0a7844" Mar 2 14:28:54.591889 containerd[1574]: time="2026-03-02T14:28:54.586259756Z" level=error msg="Failed to destroy network for sandbox \"aa3623815a8b36fa3429cb10cfbdf9d71cd9a529350a760b35d2c6a73f27eefe\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.615895 containerd[1574]: time="2026-03-02T14:28:54.613730070Z" level=error msg="Failed to destroy network for sandbox \"62a71f690cb007e9d08b90a552126930adb25b791e9091fe0db4901d1cc6553a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.630507 containerd[1574]: time="2026-03-02T14:28:54.630193723Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d7f6b6d6-zx66p,Uid:a5277f53-71c0-4d52-bac9-fd4719f10daa,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa3623815a8b36fa3429cb10cfbdf9d71cd9a529350a760b35d2c6a73f27eefe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.634506 kubelet[2823]: E0302 14:28:54.633234 2823 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa3623815a8b36fa3429cb10cfbdf9d71cd9a529350a760b35d2c6a73f27eefe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.634506 kubelet[2823]: E0302 14:28:54.633318 2823 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa3623815a8b36fa3429cb10cfbdf9d71cd9a529350a760b35d2c6a73f27eefe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d7f6b6d6-zx66p" Mar 2 14:28:54.634506 kubelet[2823]: E0302 14:28:54.633341 2823 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa3623815a8b36fa3429cb10cfbdf9d71cd9a529350a760b35d2c6a73f27eefe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d7f6b6d6-zx66p" Mar 2 14:28:54.634659 kubelet[2823]: E0302 14:28:54.633399 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d7f6b6d6-zx66p_calico-system(a5277f53-71c0-4d52-bac9-fd4719f10daa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d7f6b6d6-zx66p_calico-system(a5277f53-71c0-4d52-bac9-fd4719f10daa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"aa3623815a8b36fa3429cb10cfbdf9d71cd9a529350a760b35d2c6a73f27eefe\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d7f6b6d6-zx66p" podUID="a5277f53-71c0-4d52-bac9-fd4719f10daa" Mar 2 14:28:54.639077 containerd[1574]: time="2026-03-02T14:28:54.638619730Z" level=error msg="Failed to destroy network for sandbox \"df99d2bfe507e89a594684b8c1f3f295f80b05b5eb93019392f7882b1d5779d1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.640673 containerd[1574]: time="2026-03-02T14:28:54.640554220Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68647fc8d9-mvfxc,Uid:0dd516cb-8822-4937-b80a-6a7395992c08,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"62a71f690cb007e9d08b90a552126930adb25b791e9091fe0db4901d1cc6553a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.643316 kubelet[2823]: E0302 14:28:54.643055 2823 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"62a71f690cb007e9d08b90a552126930adb25b791e9091fe0db4901d1cc6553a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.643316 kubelet[2823]: E0302 14:28:54.643104 2823 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"62a71f690cb007e9d08b90a552126930adb25b791e9091fe0db4901d1cc6553a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-68647fc8d9-mvfxc" Mar 2 14:28:54.643316 kubelet[2823]: E0302 14:28:54.643133 2823 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"62a71f690cb007e9d08b90a552126930adb25b791e9091fe0db4901d1cc6553a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-68647fc8d9-mvfxc" Mar 2 14:28:54.643524 kubelet[2823]: E0302 14:28:54.643214 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-68647fc8d9-mvfxc_calico-system(0dd516cb-8822-4937-b80a-6a7395992c08)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-68647fc8d9-mvfxc_calico-system(0dd516cb-8822-4937-b80a-6a7395992c08)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"62a71f690cb007e9d08b90a552126930adb25b791e9091fe0db4901d1cc6553a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-68647fc8d9-mvfxc" podUID="0dd516cb-8822-4937-b80a-6a7395992c08" Mar 2 14:28:54.646122 containerd[1574]: time="2026-03-02T14:28:54.646039234Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-xrdgw,Uid:71148f0a-7223-4509-99b6-37365349cb1b,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"df99d2bfe507e89a594684b8c1f3f295f80b05b5eb93019392f7882b1d5779d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.648775 kubelet[2823]: E0302 14:28:54.648075 2823 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df99d2bfe507e89a594684b8c1f3f295f80b05b5eb93019392f7882b1d5779d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.648775 kubelet[2823]: E0302 14:28:54.648122 2823 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df99d2bfe507e89a594684b8c1f3f295f80b05b5eb93019392f7882b1d5779d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-xrdgw" Mar 2 14:28:54.648775 kubelet[2823]: E0302 14:28:54.648146 2823 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df99d2bfe507e89a594684b8c1f3f295f80b05b5eb93019392f7882b1d5779d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-xrdgw" Mar 2 14:28:54.648945 kubelet[2823]: E0302 14:28:54.648196 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-xrdgw_kube-system(71148f0a-7223-4509-99b6-37365349cb1b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-xrdgw_kube-system(71148f0a-7223-4509-99b6-37365349cb1b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"df99d2bfe507e89a594684b8c1f3f295f80b05b5eb93019392f7882b1d5779d1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-xrdgw" podUID="71148f0a-7223-4509-99b6-37365349cb1b" Mar 2 14:28:54.773600 containerd[1574]: time="2026-03-02T14:28:54.773547989Z" level=error msg="Failed to destroy network for sandbox \"e8da7ce1343f5cd98bac6262a1ab8a4bbbb0a7aeb7b624488ef4d737857684a6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.790923 containerd[1574]: time="2026-03-02T14:28:54.790786816Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fgxfk,Uid:f81f0355-d756-4ec2-b9d0-9639b176b2ea,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8da7ce1343f5cd98bac6262a1ab8a4bbbb0a7aeb7b624488ef4d737857684a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.791672 kubelet[2823]: E0302 14:28:54.791257 2823 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8da7ce1343f5cd98bac6262a1ab8a4bbbb0a7aeb7b624488ef4d737857684a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 14:28:54.791672 kubelet[2823]: E0302 14:28:54.791326 2823 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8da7ce1343f5cd98bac6262a1ab8a4bbbb0a7aeb7b624488ef4d737857684a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-fgxfk" Mar 2 14:28:54.791672 kubelet[2823]: E0302 14:28:54.791349 2823 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8da7ce1343f5cd98bac6262a1ab8a4bbbb0a7aeb7b624488ef4d737857684a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-fgxfk" Mar 2 14:28:54.792155 kubelet[2823]: E0302 14:28:54.791407 2823 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-fgxfk_calico-system(f81f0355-d756-4ec2-b9d0-9639b176b2ea)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-fgxfk_calico-system(f81f0355-d756-4ec2-b9d0-9639b176b2ea)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e8da7ce1343f5cd98bac6262a1ab8a4bbbb0a7aeb7b624488ef4d737857684a6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-fgxfk" podUID="f81f0355-d756-4ec2-b9d0-9639b176b2ea" Mar 2 14:28:55.129963 kubelet[2823]: I0302 14:28:55.127992 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-mbhdq" podStartSLOduration=12.423414525 podStartE2EDuration="1m9.127974228s" podCreationTimestamp="2026-03-02 14:27:46 +0000 UTC" firstStartedPulling="2026-03-02 14:27:51.15164863 +0000 UTC m=+34.585316633" lastFinishedPulling="2026-03-02 14:28:47.856208333 +0000 UTC m=+91.289876336" observedRunningTime="2026-03-02 14:28:55.113967954 +0000 UTC m=+98.547635988" watchObservedRunningTime="2026-03-02 14:28:55.127974228 +0000 UTC m=+98.561642230" Mar 2 14:28:55.291615 systemd[1]: run-netns-cni\x2d06092c7a\x2d0324\x2d79b8\x2d2f95\x2de13a6f36c843.mount: Deactivated successfully. Mar 2 14:28:55.291771 systemd[1]: run-netns-cni\x2db326d779\x2d89f9\x2df111\x2db6b1\x2d54f9e1dc156b.mount: Deactivated successfully. Mar 2 14:28:55.291930 systemd[1]: run-netns-cni\x2d03939a32\x2d4d06\x2dbb8a\x2d811b\x2dcfb059a65e05.mount: Deactivated successfully. Mar 2 14:28:55.292020 systemd[1]: run-netns-cni\x2d4890b8d7\x2de14e\x2d1b30\x2d979a\x2dc6dfe63c8385.mount: Deactivated successfully. Mar 2 14:28:55.668667 kubelet[2823]: I0302 14:28:55.668612 2823 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1acbed91-a452-4e11-a018-a8670f52ea1a-whisker-ca-bundle\") pod \"1acbed91-a452-4e11-a018-a8670f52ea1a\" (UID: \"1acbed91-a452-4e11-a018-a8670f52ea1a\") " Mar 2 14:28:55.672260 kubelet[2823]: I0302 14:28:55.669540 2823 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/1acbed91-a452-4e11-a018-a8670f52ea1a-whisker-backend-key-pair\") pod \"1acbed91-a452-4e11-a018-a8670f52ea1a\" (UID: \"1acbed91-a452-4e11-a018-a8670f52ea1a\") " Mar 2 14:28:55.672498 kubelet[2823]: I0302 14:28:55.672427 2823 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvpfh\" (UniqueName: \"kubernetes.io/projected/1acbed91-a452-4e11-a018-a8670f52ea1a-kube-api-access-bvpfh\") pod \"1acbed91-a452-4e11-a018-a8670f52ea1a\" (UID: \"1acbed91-a452-4e11-a018-a8670f52ea1a\") " Mar 2 14:28:55.672711 kubelet[2823]: I0302 14:28:55.672690 2823 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/1acbed91-a452-4e11-a018-a8670f52ea1a-nginx-config\") pod \"1acbed91-a452-4e11-a018-a8670f52ea1a\" (UID: \"1acbed91-a452-4e11-a018-a8670f52ea1a\") " Mar 2 14:28:55.674496 kubelet[2823]: I0302 14:28:55.674418 2823 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1acbed91-a452-4e11-a018-a8670f52ea1a-nginx-config" (OuterVolumeSpecName: "nginx-config") pod "1acbed91-a452-4e11-a018-a8670f52ea1a" (UID: "1acbed91-a452-4e11-a018-a8670f52ea1a"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 2 14:28:55.675517 kubelet[2823]: I0302 14:28:55.675487 2823 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1acbed91-a452-4e11-a018-a8670f52ea1a-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "1acbed91-a452-4e11-a018-a8670f52ea1a" (UID: "1acbed91-a452-4e11-a018-a8670f52ea1a"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 2 14:28:55.702727 systemd[1]: var-lib-kubelet-pods-1acbed91\x2da452\x2d4e11\x2da018\x2da8670f52ea1a-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Mar 2 14:28:55.704007 kubelet[2823]: I0302 14:28:55.703253 2823 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1acbed91-a452-4e11-a018-a8670f52ea1a-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "1acbed91-a452-4e11-a018-a8670f52ea1a" (UID: "1acbed91-a452-4e11-a018-a8670f52ea1a"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 2 14:28:55.714187 kubelet[2823]: I0302 14:28:55.714092 2823 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1acbed91-a452-4e11-a018-a8670f52ea1a-kube-api-access-bvpfh" (OuterVolumeSpecName: "kube-api-access-bvpfh") pod "1acbed91-a452-4e11-a018-a8670f52ea1a" (UID: "1acbed91-a452-4e11-a018-a8670f52ea1a"). InnerVolumeSpecName "kube-api-access-bvpfh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 14:28:55.718209 systemd[1]: var-lib-kubelet-pods-1acbed91\x2da452\x2d4e11\x2da018\x2da8670f52ea1a-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dbvpfh.mount: Deactivated successfully. Mar 2 14:28:55.775281 kubelet[2823]: I0302 14:28:55.775182 2823 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/1acbed91-a452-4e11-a018-a8670f52ea1a-nginx-config\") on node \"localhost\" DevicePath \"\"" Mar 2 14:28:55.775281 kubelet[2823]: I0302 14:28:55.775225 2823 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1acbed91-a452-4e11-a018-a8670f52ea1a-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Mar 2 14:28:55.775281 kubelet[2823]: I0302 14:28:55.775239 2823 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/1acbed91-a452-4e11-a018-a8670f52ea1a-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Mar 2 14:28:55.775281 kubelet[2823]: I0302 14:28:55.775251 2823 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-bvpfh\" (UniqueName: \"kubernetes.io/projected/1acbed91-a452-4e11-a018-a8670f52ea1a-kube-api-access-bvpfh\") on node \"localhost\" DevicePath \"\"" Mar 2 14:28:56.063242 systemd[1]: Removed slice kubepods-besteffort-pod1acbed91_a452_4e11_a018_a8670f52ea1a.slice - libcontainer container kubepods-besteffort-pod1acbed91_a452_4e11_a018_a8670f52ea1a.slice. Mar 2 14:28:56.612154 systemd[1]: Created slice kubepods-besteffort-poda99551fb_e1b1_42f4_8be2_271fd4e2a6ee.slice - libcontainer container kubepods-besteffort-poda99551fb_e1b1_42f4_8be2_271fd4e2a6ee.slice. Mar 2 14:28:56.715983 kubelet[2823]: I0302 14:28:56.711067 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/a99551fb-e1b1-42f4-8be2-271fd4e2a6ee-whisker-backend-key-pair\") pod \"whisker-7866c7fc4c-w9kss\" (UID: \"a99551fb-e1b1-42f4-8be2-271fd4e2a6ee\") " pod="calico-system/whisker-7866c7fc4c-w9kss" Mar 2 14:28:56.715983 kubelet[2823]: I0302 14:28:56.711156 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/a99551fb-e1b1-42f4-8be2-271fd4e2a6ee-nginx-config\") pod \"whisker-7866c7fc4c-w9kss\" (UID: \"a99551fb-e1b1-42f4-8be2-271fd4e2a6ee\") " pod="calico-system/whisker-7866c7fc4c-w9kss" Mar 2 14:28:56.715983 kubelet[2823]: I0302 14:28:56.711192 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a99551fb-e1b1-42f4-8be2-271fd4e2a6ee-whisker-ca-bundle\") pod \"whisker-7866c7fc4c-w9kss\" (UID: \"a99551fb-e1b1-42f4-8be2-271fd4e2a6ee\") " pod="calico-system/whisker-7866c7fc4c-w9kss" Mar 2 14:28:56.715983 kubelet[2823]: I0302 14:28:56.711211 2823 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7t9c\" (UniqueName: \"kubernetes.io/projected/a99551fb-e1b1-42f4-8be2-271fd4e2a6ee-kube-api-access-r7t9c\") pod \"whisker-7866c7fc4c-w9kss\" (UID: \"a99551fb-e1b1-42f4-8be2-271fd4e2a6ee\") " pod="calico-system/whisker-7866c7fc4c-w9kss" Mar 2 14:28:56.962240 containerd[1574]: time="2026-03-02T14:28:56.961728698Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7866c7fc4c-w9kss,Uid:a99551fb-e1b1-42f4-8be2-271fd4e2a6ee,Namespace:calico-system,Attempt:0,}" Mar 2 14:28:57.120475 kubelet[2823]: I0302 14:28:57.120375 2823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1acbed91-a452-4e11-a018-a8670f52ea1a" path="/var/lib/kubelet/pods/1acbed91-a452-4e11-a018-a8670f52ea1a/volumes" Mar 2 14:28:57.958082 systemd-networkd[1473]: cali5fb723755d9: Link UP Mar 2 14:28:57.958665 systemd-networkd[1473]: cali5fb723755d9: Gained carrier Mar 2 14:28:58.044117 containerd[1574]: 2026-03-02 14:28:57.116 [ERROR][4170] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Mar 2 14:28:58.044117 containerd[1574]: 2026-03-02 14:28:57.345 [INFO][4170] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--7866c7fc4c--w9kss-eth0 whisker-7866c7fc4c- calico-system a99551fb-e1b1-42f4-8be2-271fd4e2a6ee 1100 0 2026-03-02 14:28:56 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:7866c7fc4c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-7866c7fc4c-w9kss eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali5fb723755d9 [] [] }} ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Namespace="calico-system" Pod="whisker-7866c7fc4c-w9kss" WorkloadEndpoint="localhost-k8s-whisker--7866c7fc4c--w9kss-" Mar 2 14:28:58.044117 containerd[1574]: 2026-03-02 14:28:57.346 [INFO][4170] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Namespace="calico-system" Pod="whisker-7866c7fc4c-w9kss" WorkloadEndpoint="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" Mar 2 14:28:58.044117 containerd[1574]: 2026-03-02 14:28:57.534 [INFO][4184] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" HandleID="k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Workload="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.568 [INFO][4184] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" HandleID="k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Workload="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fc30), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-7866c7fc4c-w9kss", "timestamp":"2026-03-02 14:28:57.534014988 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000378420)} Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.568 [INFO][4184] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.568 [INFO][4184] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.568 [INFO][4184] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.602 [INFO][4184] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" host="localhost" Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.665 [INFO][4184] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.716 [INFO][4184] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.733 [INFO][4184] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.760 [INFO][4184] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 2 14:28:58.044967 containerd[1574]: 2026-03-02 14:28:57.760 [INFO][4184] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" host="localhost" Mar 2 14:28:58.045382 containerd[1574]: 2026-03-02 14:28:57.775 [INFO][4184] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c Mar 2 14:28:58.045382 containerd[1574]: 2026-03-02 14:28:57.807 [INFO][4184] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" host="localhost" Mar 2 14:28:58.045382 containerd[1574]: 2026-03-02 14:28:57.865 [INFO][4184] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" host="localhost" Mar 2 14:28:58.045382 containerd[1574]: 2026-03-02 14:28:57.865 [INFO][4184] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" host="localhost" Mar 2 14:28:58.045382 containerd[1574]: 2026-03-02 14:28:57.865 [INFO][4184] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 14:28:58.045382 containerd[1574]: 2026-03-02 14:28:57.865 [INFO][4184] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" HandleID="k8s-pod-network.8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Workload="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" Mar 2 14:28:58.045679 containerd[1574]: 2026-03-02 14:28:57.883 [INFO][4170] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Namespace="calico-system" Pod="whisker-7866c7fc4c-w9kss" WorkloadEndpoint="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--7866c7fc4c--w9kss-eth0", GenerateName:"whisker-7866c7fc4c-", Namespace:"calico-system", SelfLink:"", UID:"a99551fb-e1b1-42f4-8be2-271fd4e2a6ee", ResourceVersion:"1100", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 28, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7866c7fc4c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-7866c7fc4c-w9kss", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali5fb723755d9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:28:58.045679 containerd[1574]: 2026-03-02 14:28:57.885 [INFO][4170] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Namespace="calico-system" Pod="whisker-7866c7fc4c-w9kss" WorkloadEndpoint="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" Mar 2 14:28:58.054963 containerd[1574]: 2026-03-02 14:28:57.885 [INFO][4170] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5fb723755d9 ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Namespace="calico-system" Pod="whisker-7866c7fc4c-w9kss" WorkloadEndpoint="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" Mar 2 14:28:58.054963 containerd[1574]: 2026-03-02 14:28:57.963 [INFO][4170] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Namespace="calico-system" Pod="whisker-7866c7fc4c-w9kss" WorkloadEndpoint="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" Mar 2 14:28:58.055039 containerd[1574]: 2026-03-02 14:28:57.963 [INFO][4170] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Namespace="calico-system" Pod="whisker-7866c7fc4c-w9kss" WorkloadEndpoint="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--7866c7fc4c--w9kss-eth0", GenerateName:"whisker-7866c7fc4c-", Namespace:"calico-system", SelfLink:"", UID:"a99551fb-e1b1-42f4-8be2-271fd4e2a6ee", ResourceVersion:"1100", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 28, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7866c7fc4c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c", Pod:"whisker-7866c7fc4c-w9kss", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali5fb723755d9", MAC:"82:f0:f9:83:28:2e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:28:58.055164 containerd[1574]: 2026-03-02 14:28:58.025 [INFO][4170] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" Namespace="calico-system" Pod="whisker-7866c7fc4c-w9kss" WorkloadEndpoint="localhost-k8s-whisker--7866c7fc4c--w9kss-eth0" Mar 2 14:28:58.417736 containerd[1574]: time="2026-03-02T14:28:58.416895385Z" level=info msg="connecting to shim 8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c" address="unix:///run/containerd/s/84ca8c0bff80fc9baeb27788fdf09e0e9a7d5f5761a234a0404dda2f4f7ca2da" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:28:58.525110 systemd[1]: Started cri-containerd-8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c.scope - libcontainer container 8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c. Mar 2 14:28:58.661226 systemd-resolved[1474]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 14:28:58.911678 containerd[1574]: time="2026-03-02T14:28:58.907529850Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7866c7fc4c-w9kss,Uid:a99551fb-e1b1-42f4-8be2-271fd4e2a6ee,Namespace:calico-system,Attempt:0,} returns sandbox id \"8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c\"" Mar 2 14:28:58.911678 containerd[1574]: time="2026-03-02T14:28:58.910658731Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.3\"" Mar 2 14:28:59.859528 systemd-networkd[1473]: cali5fb723755d9: Gained IPv6LL Mar 2 14:29:01.212251 containerd[1574]: time="2026-03-02T14:29:01.206148331Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:01.218233 containerd[1574]: time="2026-03-02T14:29:01.217774321Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.3: active requests=0, bytes read=6036825" Mar 2 14:29:01.232209 containerd[1574]: time="2026-03-02T14:29:01.226718573Z" level=info msg="ImageCreate event name:\"sha256:a4bcedf3b244f5fd0077952f436fd9486e0e6b974a358c85a962b60303e94c02\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:01.248919 containerd[1574]: time="2026-03-02T14:29:01.247084674Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:3a388b567fff5cc31c64399d4af0fd03d2f4d243ef26e6f6b77a49386dbadeca\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:01.264322 containerd[1574]: time="2026-03-02T14:29:01.263626787Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.3\" with image id \"sha256:a4bcedf3b244f5fd0077952f436fd9486e0e6b974a358c85a962b60303e94c02\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:3a388b567fff5cc31c64399d4af0fd03d2f4d243ef26e6f6b77a49386dbadeca\", size \"7592862\" in 2.352931928s" Mar 2 14:29:01.264322 containerd[1574]: time="2026-03-02T14:29:01.263876553Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.3\" returns image reference \"sha256:a4bcedf3b244f5fd0077952f436fd9486e0e6b974a358c85a962b60303e94c02\"" Mar 2 14:29:01.327010 containerd[1574]: time="2026-03-02T14:29:01.325428161Z" level=info msg="CreateContainer within sandbox \"8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Mar 2 14:29:01.477586 containerd[1574]: time="2026-03-02T14:29:01.474144809Z" level=info msg="Container b23179b718093269e29b841fce79d65628829db14d6b5709da60d06d75e15045: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:01.560234 containerd[1574]: time="2026-03-02T14:29:01.557880167Z" level=info msg="CreateContainer within sandbox \"8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"b23179b718093269e29b841fce79d65628829db14d6b5709da60d06d75e15045\"" Mar 2 14:29:01.560234 containerd[1574]: time="2026-03-02T14:29:01.558885319Z" level=info msg="StartContainer for \"b23179b718093269e29b841fce79d65628829db14d6b5709da60d06d75e15045\"" Mar 2 14:29:01.566386 containerd[1574]: time="2026-03-02T14:29:01.566294560Z" level=info msg="connecting to shim b23179b718093269e29b841fce79d65628829db14d6b5709da60d06d75e15045" address="unix:///run/containerd/s/84ca8c0bff80fc9baeb27788fdf09e0e9a7d5f5761a234a0404dda2f4f7ca2da" protocol=ttrpc version=3 Mar 2 14:29:01.723795 systemd[1]: Started cri-containerd-b23179b718093269e29b841fce79d65628829db14d6b5709da60d06d75e15045.scope - libcontainer container b23179b718093269e29b841fce79d65628829db14d6b5709da60d06d75e15045. Mar 2 14:29:02.022503 containerd[1574]: time="2026-03-02T14:29:02.020415223Z" level=info msg="StartContainer for \"b23179b718093269e29b841fce79d65628829db14d6b5709da60d06d75e15045\" returns successfully" Mar 2 14:29:02.044697 containerd[1574]: time="2026-03-02T14:29:02.044618227Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\"" Mar 2 14:29:02.822641 systemd-networkd[1473]: vxlan.calico: Link UP Mar 2 14:29:02.822654 systemd-networkd[1473]: vxlan.calico: Gained carrier Mar 2 14:29:04.275976 systemd-networkd[1473]: vxlan.calico: Gained IPv6LL Mar 2 14:29:05.126679 containerd[1574]: time="2026-03-02T14:29:05.126631049Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6d7c47ccd9-vmfz4,Uid:7633016e-887b-423d-9baf-5d37aa0a7844,Namespace:calico-system,Attempt:0,}" Mar 2 14:29:05.796554 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount694807234.mount: Deactivated successfully. Mar 2 14:29:05.947247 containerd[1574]: time="2026-03-02T14:29:05.947107776Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:05.964632 containerd[1574]: time="2026-03-02T14:29:05.964571844Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.3: active requests=0, bytes read=17599119" Mar 2 14:29:05.982758 containerd[1574]: time="2026-03-02T14:29:05.978636451Z" level=info msg="ImageCreate event name:\"sha256:fd911f8f9ea58b19b827b1f51a4c19e899291759aca4ed03c388788897668b8f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:05.980714 systemd-networkd[1473]: cali51e3ce39796: Link UP Mar 2 14:29:05.981082 systemd-networkd[1473]: cali51e3ce39796: Gained carrier Mar 2 14:29:05.992044 containerd[1574]: time="2026-03-02T14:29:05.992005854Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:359cb5c751e049ac0bb62c4f7e49b1ac81c59935c70715f5ff4c39a757bf9f38\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:05.996141 containerd[1574]: time="2026-03-02T14:29:05.995959597Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\" with image id \"sha256:fd911f8f9ea58b19b827b1f51a4c19e899291759aca4ed03c388788897668b8f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:359cb5c751e049ac0bb62c4f7e49b1ac81c59935c70715f5ff4c39a757bf9f38\", size \"17598949\" in 3.951265017s" Mar 2 14:29:05.996141 containerd[1574]: time="2026-03-02T14:29:05.996034136Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\" returns image reference \"sha256:fd911f8f9ea58b19b827b1f51a4c19e899291759aca4ed03c388788897668b8f\"" Mar 2 14:29:06.015211 containerd[1574]: time="2026-03-02T14:29:06.015100623Z" level=info msg="CreateContainer within sandbox \"8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Mar 2 14:29:06.077190 containerd[1574]: time="2026-03-02T14:29:06.074271039Z" level=info msg="Container f6a69308fe9487518c6cf97dcca9cdbf57359fc80e3f0efdbef912b3fe6ea96b: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:06.079322 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2090715000.mount: Deactivated successfully. Mar 2 14:29:06.118533 containerd[1574]: 2026-03-02 14:29:05.482 [INFO][4530] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0 calico-kube-controllers-6d7c47ccd9- calico-system 7633016e-887b-423d-9baf-5d37aa0a7844 1042 0 2026-03-02 14:27:50 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6d7c47ccd9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-6d7c47ccd9-vmfz4 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali51e3ce39796 [] [] }} ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Namespace="calico-system" Pod="calico-kube-controllers-6d7c47ccd9-vmfz4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-" Mar 2 14:29:06.118533 containerd[1574]: 2026-03-02 14:29:05.482 [INFO][4530] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Namespace="calico-system" Pod="calico-kube-controllers-6d7c47ccd9-vmfz4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" Mar 2 14:29:06.118533 containerd[1574]: 2026-03-02 14:29:05.671 [INFO][4544] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" HandleID="k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Workload="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.705 [INFO][4544] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" HandleID="k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Workload="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000375d10), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-6d7c47ccd9-vmfz4", "timestamp":"2026-03-02 14:29:05.671163837 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0004151e0)} Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.706 [INFO][4544] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.706 [INFO][4544] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.706 [INFO][4544] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.738 [INFO][4544] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" host="localhost" Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.774 [INFO][4544] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.831 [INFO][4544] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.847 [INFO][4544] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:06.119236 containerd[1574]: 2026-03-02 14:29:05.866 [INFO][4544] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:06.120327 containerd[1574]: 2026-03-02 14:29:05.866 [INFO][4544] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" host="localhost" Mar 2 14:29:06.120327 containerd[1574]: 2026-03-02 14:29:05.886 [INFO][4544] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc Mar 2 14:29:06.120327 containerd[1574]: 2026-03-02 14:29:05.929 [INFO][4544] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" host="localhost" Mar 2 14:29:06.120327 containerd[1574]: 2026-03-02 14:29:05.960 [INFO][4544] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" host="localhost" Mar 2 14:29:06.120327 containerd[1574]: 2026-03-02 14:29:05.960 [INFO][4544] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" host="localhost" Mar 2 14:29:06.120327 containerd[1574]: 2026-03-02 14:29:05.960 [INFO][4544] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 14:29:06.120327 containerd[1574]: 2026-03-02 14:29:05.960 [INFO][4544] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" HandleID="k8s-pod-network.5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Workload="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" Mar 2 14:29:06.120637 containerd[1574]: 2026-03-02 14:29:05.968 [INFO][4530] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Namespace="calico-system" Pod="calico-kube-controllers-6d7c47ccd9-vmfz4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0", GenerateName:"calico-kube-controllers-6d7c47ccd9-", Namespace:"calico-system", SelfLink:"", UID:"7633016e-887b-423d-9baf-5d37aa0a7844", ResourceVersion:"1042", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6d7c47ccd9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-6d7c47ccd9-vmfz4", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali51e3ce39796", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:06.120786 containerd[1574]: 2026-03-02 14:29:05.972 [INFO][4530] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Namespace="calico-system" Pod="calico-kube-controllers-6d7c47ccd9-vmfz4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" Mar 2 14:29:06.120786 containerd[1574]: 2026-03-02 14:29:05.972 [INFO][4530] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali51e3ce39796 ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Namespace="calico-system" Pod="calico-kube-controllers-6d7c47ccd9-vmfz4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" Mar 2 14:29:06.120786 containerd[1574]: 2026-03-02 14:29:05.992 [INFO][4530] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Namespace="calico-system" Pod="calico-kube-controllers-6d7c47ccd9-vmfz4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" Mar 2 14:29:06.121046 containerd[1574]: 2026-03-02 14:29:05.992 [INFO][4530] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Namespace="calico-system" Pod="calico-kube-controllers-6d7c47ccd9-vmfz4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0", GenerateName:"calico-kube-controllers-6d7c47ccd9-", Namespace:"calico-system", SelfLink:"", UID:"7633016e-887b-423d-9baf-5d37aa0a7844", ResourceVersion:"1042", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6d7c47ccd9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc", Pod:"calico-kube-controllers-6d7c47ccd9-vmfz4", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali51e3ce39796", MAC:"52:15:3b:ba:fb:c9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:06.121171 containerd[1574]: 2026-03-02 14:29:06.088 [INFO][4530] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" Namespace="calico-system" Pod="calico-kube-controllers-6d7c47ccd9-vmfz4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6d7c47ccd9--vmfz4-eth0" Mar 2 14:29:06.173215 containerd[1574]: time="2026-03-02T14:29:06.172973103Z" level=info msg="CreateContainer within sandbox \"8d89049ebe7d6194e2fac6a38378ca4a588effda02b7218ee64ea41415cf332c\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"f6a69308fe9487518c6cf97dcca9cdbf57359fc80e3f0efdbef912b3fe6ea96b\"" Mar 2 14:29:06.180495 containerd[1574]: time="2026-03-02T14:29:06.180371454Z" level=info msg="StartContainer for \"f6a69308fe9487518c6cf97dcca9cdbf57359fc80e3f0efdbef912b3fe6ea96b\"" Mar 2 14:29:06.243604 containerd[1574]: time="2026-03-02T14:29:06.243492114Z" level=info msg="connecting to shim f6a69308fe9487518c6cf97dcca9cdbf57359fc80e3f0efdbef912b3fe6ea96b" address="unix:///run/containerd/s/84ca8c0bff80fc9baeb27788fdf09e0e9a7d5f5761a234a0404dda2f4f7ca2da" protocol=ttrpc version=3 Mar 2 14:29:06.337641 containerd[1574]: time="2026-03-02T14:29:06.337397964Z" level=info msg="connecting to shim 5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc" address="unix:///run/containerd/s/1a6e4873526a1d1e5d4fc85305ba66537be9e818ce17d96defc3c65d056b3dc2" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:29:06.413079 systemd[1]: Started cri-containerd-f6a69308fe9487518c6cf97dcca9cdbf57359fc80e3f0efdbef912b3fe6ea96b.scope - libcontainer container f6a69308fe9487518c6cf97dcca9cdbf57359fc80e3f0efdbef912b3fe6ea96b. Mar 2 14:29:06.535970 systemd[1]: Started cri-containerd-5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc.scope - libcontainer container 5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc. Mar 2 14:29:06.630874 systemd-resolved[1474]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 14:29:06.724973 containerd[1574]: time="2026-03-02T14:29:06.722166864Z" level=info msg="StartContainer for \"f6a69308fe9487518c6cf97dcca9cdbf57359fc80e3f0efdbef912b3fe6ea96b\" returns successfully" Mar 2 14:29:06.835920 containerd[1574]: time="2026-03-02T14:29:06.830083522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6d7c47ccd9-vmfz4,Uid:7633016e-887b-423d-9baf-5d37aa0a7844,Namespace:calico-system,Attempt:0,} returns sandbox id \"5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc\"" Mar 2 14:29:06.842964 containerd[1574]: time="2026-03-02T14:29:06.841259525Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\"" Mar 2 14:29:07.108493 containerd[1574]: time="2026-03-02T14:29:07.108356553Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d7f6b6d6-zx66p,Uid:a5277f53-71c0-4d52-bac9-fd4719f10daa,Namespace:calico-system,Attempt:0,}" Mar 2 14:29:07.412770 kubelet[2823]: I0302 14:29:07.412173 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-7866c7fc4c-w9kss" podStartSLOduration=4.320790349 podStartE2EDuration="11.412060278s" podCreationTimestamp="2026-03-02 14:28:56 +0000 UTC" firstStartedPulling="2026-03-02 14:28:58.910125007 +0000 UTC m=+102.343793010" lastFinishedPulling="2026-03-02 14:29:06.001394937 +0000 UTC m=+109.435062939" observedRunningTime="2026-03-02 14:29:07.397416968 +0000 UTC m=+110.831085002" watchObservedRunningTime="2026-03-02 14:29:07.412060278 +0000 UTC m=+110.845728301" Mar 2 14:29:07.472614 systemd-networkd[1473]: cali51e3ce39796: Gained IPv6LL Mar 2 14:29:07.939402 systemd-networkd[1473]: cali3a5cdfac3c3: Link UP Mar 2 14:29:07.943512 systemd-networkd[1473]: cali3a5cdfac3c3: Gained carrier Mar 2 14:29:08.052025 containerd[1574]: 2026-03-02 14:29:07.343 [INFO][4652] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0 goldmane-54d7f6b6d6- calico-system a5277f53-71c0-4d52-bac9-fd4719f10daa 1052 0 2026-03-02 14:27:44 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d7f6b6d6 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-54d7f6b6d6-zx66p eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali3a5cdfac3c3 [] [] }} ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Namespace="calico-system" Pod="goldmane-54d7f6b6d6-zx66p" WorkloadEndpoint="localhost-k8s-goldmane--54d7f6b6d6--zx66p-" Mar 2 14:29:08.052025 containerd[1574]: 2026-03-02 14:29:07.343 [INFO][4652] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Namespace="calico-system" Pod="goldmane-54d7f6b6d6-zx66p" WorkloadEndpoint="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" Mar 2 14:29:08.052025 containerd[1574]: 2026-03-02 14:29:07.634 [INFO][4667] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" HandleID="k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Workload="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.654 [INFO][4667] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" HandleID="k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Workload="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0006e4620), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-54d7f6b6d6-zx66p", "timestamp":"2026-03-02 14:29:07.634980962 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000216420)} Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.654 [INFO][4667] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.655 [INFO][4667] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.655 [INFO][4667] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.668 [INFO][4667] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" host="localhost" Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.766 [INFO][4667] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.820 [INFO][4667] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.828 [INFO][4667] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.835 [INFO][4667] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:08.053165 containerd[1574]: 2026-03-02 14:29:07.835 [INFO][4667] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" host="localhost" Mar 2 14:29:08.058628 containerd[1574]: 2026-03-02 14:29:07.844 [INFO][4667] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec Mar 2 14:29:08.058628 containerd[1574]: 2026-03-02 14:29:07.884 [INFO][4667] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" host="localhost" Mar 2 14:29:08.058628 containerd[1574]: 2026-03-02 14:29:07.926 [INFO][4667] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" host="localhost" Mar 2 14:29:08.058628 containerd[1574]: 2026-03-02 14:29:07.926 [INFO][4667] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" host="localhost" Mar 2 14:29:08.058628 containerd[1574]: 2026-03-02 14:29:07.926 [INFO][4667] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 14:29:08.058628 containerd[1574]: 2026-03-02 14:29:07.926 [INFO][4667] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" HandleID="k8s-pod-network.a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Workload="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" Mar 2 14:29:08.058952 containerd[1574]: 2026-03-02 14:29:07.933 [INFO][4652] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Namespace="calico-system" Pod="goldmane-54d7f6b6d6-zx66p" WorkloadEndpoint="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0", GenerateName:"goldmane-54d7f6b6d6-", Namespace:"calico-system", SelfLink:"", UID:"a5277f53-71c0-4d52-bac9-fd4719f10daa", ResourceVersion:"1052", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d7f6b6d6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-54d7f6b6d6-zx66p", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3a5cdfac3c3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:08.058952 containerd[1574]: 2026-03-02 14:29:07.933 [INFO][4652] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Namespace="calico-system" Pod="goldmane-54d7f6b6d6-zx66p" WorkloadEndpoint="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" Mar 2 14:29:08.059127 containerd[1574]: 2026-03-02 14:29:07.933 [INFO][4652] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3a5cdfac3c3 ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Namespace="calico-system" Pod="goldmane-54d7f6b6d6-zx66p" WorkloadEndpoint="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" Mar 2 14:29:08.059127 containerd[1574]: 2026-03-02 14:29:07.946 [INFO][4652] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Namespace="calico-system" Pod="goldmane-54d7f6b6d6-zx66p" WorkloadEndpoint="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" Mar 2 14:29:08.059184 containerd[1574]: 2026-03-02 14:29:07.950 [INFO][4652] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Namespace="calico-system" Pod="goldmane-54d7f6b6d6-zx66p" WorkloadEndpoint="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0", GenerateName:"goldmane-54d7f6b6d6-", Namespace:"calico-system", SelfLink:"", UID:"a5277f53-71c0-4d52-bac9-fd4719f10daa", ResourceVersion:"1052", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d7f6b6d6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec", Pod:"goldmane-54d7f6b6d6-zx66p", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3a5cdfac3c3", MAC:"2e:eb:a3:6a:d2:09", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:08.059273 containerd[1574]: 2026-03-02 14:29:08.017 [INFO][4652] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" Namespace="calico-system" Pod="goldmane-54d7f6b6d6-zx66p" WorkloadEndpoint="localhost-k8s-goldmane--54d7f6b6d6--zx66p-eth0" Mar 2 14:29:08.151200 containerd[1574]: time="2026-03-02T14:29:08.149953207Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fgxfk,Uid:f81f0355-d756-4ec2-b9d0-9639b176b2ea,Namespace:calico-system,Attempt:0,}" Mar 2 14:29:08.230782 containerd[1574]: time="2026-03-02T14:29:08.230651792Z" level=info msg="connecting to shim a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec" address="unix:///run/containerd/s/c22c5d2253d8d5372b7beb2dba1a21e665e1914cc786b8cee5ae88a5e7554d57" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:29:08.375062 systemd[1]: Started cri-containerd-a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec.scope - libcontainer container a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec. Mar 2 14:29:08.503991 systemd-resolved[1474]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 14:29:08.739193 containerd[1574]: time="2026-03-02T14:29:08.739106878Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d7f6b6d6-zx66p,Uid:a5277f53-71c0-4d52-bac9-fd4719f10daa,Namespace:calico-system,Attempt:0,} returns sandbox id \"a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec\"" Mar 2 14:29:08.965498 systemd-networkd[1473]: cali59f3669a390: Link UP Mar 2 14:29:08.971734 systemd-networkd[1473]: cali59f3669a390: Gained carrier Mar 2 14:29:09.045998 containerd[1574]: 2026-03-02 14:29:08.522 [INFO][4697] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--fgxfk-eth0 csi-node-driver- calico-system f81f0355-d756-4ec2-b9d0-9639b176b2ea 805 0 2026-03-02 14:27:48 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6db5596769 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-fgxfk eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali59f3669a390 [] [] }} ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Namespace="calico-system" Pod="csi-node-driver-fgxfk" WorkloadEndpoint="localhost-k8s-csi--node--driver--fgxfk-" Mar 2 14:29:09.045998 containerd[1574]: 2026-03-02 14:29:08.522 [INFO][4697] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Namespace="calico-system" Pod="csi-node-driver-fgxfk" WorkloadEndpoint="localhost-k8s-csi--node--driver--fgxfk-eth0" Mar 2 14:29:09.045998 containerd[1574]: 2026-03-02 14:29:08.647 [INFO][4754] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" HandleID="k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Workload="localhost-k8s-csi--node--driver--fgxfk-eth0" Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.683 [INFO][4754] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" HandleID="k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Workload="localhost-k8s-csi--node--driver--fgxfk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f620), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-fgxfk", "timestamp":"2026-03-02 14:29:08.647306294 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000113340)} Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.683 [INFO][4754] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.683 [INFO][4754] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.683 [INFO][4754] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.706 [INFO][4754] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" host="localhost" Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.738 [INFO][4754] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.782 [INFO][4754] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.796 [INFO][4754] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.813 [INFO][4754] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:09.046293 containerd[1574]: 2026-03-02 14:29:08.814 [INFO][4754] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" host="localhost" Mar 2 14:29:09.046737 containerd[1574]: 2026-03-02 14:29:08.846 [INFO][4754] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37 Mar 2 14:29:09.046737 containerd[1574]: 2026-03-02 14:29:08.903 [INFO][4754] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" host="localhost" Mar 2 14:29:09.046737 containerd[1574]: 2026-03-02 14:29:08.947 [INFO][4754] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" host="localhost" Mar 2 14:29:09.046737 containerd[1574]: 2026-03-02 14:29:08.947 [INFO][4754] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" host="localhost" Mar 2 14:29:09.046737 containerd[1574]: 2026-03-02 14:29:08.947 [INFO][4754] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 14:29:09.046737 containerd[1574]: 2026-03-02 14:29:08.947 [INFO][4754] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" HandleID="k8s-pod-network.a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Workload="localhost-k8s-csi--node--driver--fgxfk-eth0" Mar 2 14:29:09.046992 containerd[1574]: 2026-03-02 14:29:08.956 [INFO][4697] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Namespace="calico-system" Pod="csi-node-driver-fgxfk" WorkloadEndpoint="localhost-k8s-csi--node--driver--fgxfk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--fgxfk-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f81f0355-d756-4ec2-b9d0-9639b176b2ea", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6db5596769", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-fgxfk", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali59f3669a390", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:09.047118 containerd[1574]: 2026-03-02 14:29:08.956 [INFO][4697] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Namespace="calico-system" Pod="csi-node-driver-fgxfk" WorkloadEndpoint="localhost-k8s-csi--node--driver--fgxfk-eth0" Mar 2 14:29:09.047118 containerd[1574]: 2026-03-02 14:29:08.956 [INFO][4697] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali59f3669a390 ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Namespace="calico-system" Pod="csi-node-driver-fgxfk" WorkloadEndpoint="localhost-k8s-csi--node--driver--fgxfk-eth0" Mar 2 14:29:09.047118 containerd[1574]: 2026-03-02 14:29:08.968 [INFO][4697] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Namespace="calico-system" Pod="csi-node-driver-fgxfk" WorkloadEndpoint="localhost-k8s-csi--node--driver--fgxfk-eth0" Mar 2 14:29:09.047206 containerd[1574]: 2026-03-02 14:29:08.971 [INFO][4697] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Namespace="calico-system" Pod="csi-node-driver-fgxfk" WorkloadEndpoint="localhost-k8s-csi--node--driver--fgxfk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--fgxfk-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f81f0355-d756-4ec2-b9d0-9639b176b2ea", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6db5596769", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37", Pod:"csi-node-driver-fgxfk", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali59f3669a390", MAC:"be:52:fd:c9:9c:14", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:09.047315 containerd[1574]: 2026-03-02 14:29:09.031 [INFO][4697] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" Namespace="calico-system" Pod="csi-node-driver-fgxfk" WorkloadEndpoint="localhost-k8s-csi--node--driver--fgxfk-eth0" Mar 2 14:29:09.130324 containerd[1574]: time="2026-03-02T14:29:09.130079206Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68647fc8d9-p874s,Uid:ceacecf8-f59e-4263-8232-2bc003edccd4,Namespace:calico-system,Attempt:0,}" Mar 2 14:29:09.143249 kubelet[2823]: E0302 14:29:09.143169 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:09.156028 containerd[1574]: time="2026-03-02T14:29:09.145748898Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-xlqmd,Uid:bc8db32a-755c-44fc-abe9-e927df38b7b5,Namespace:kube-system,Attempt:0,}" Mar 2 14:29:09.156028 containerd[1574]: time="2026-03-02T14:29:09.155641991Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68647fc8d9-mvfxc,Uid:0dd516cb-8822-4937-b80a-6a7395992c08,Namespace:calico-system,Attempt:0,}" Mar 2 14:29:09.165000 kubelet[2823]: E0302 14:29:09.164801 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:09.169765 containerd[1574]: time="2026-03-02T14:29:09.169553345Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-xrdgw,Uid:71148f0a-7223-4509-99b6-37365349cb1b,Namespace:kube-system,Attempt:0,}" Mar 2 14:29:09.264090 systemd-networkd[1473]: cali3a5cdfac3c3: Gained IPv6LL Mar 2 14:29:09.397365 containerd[1574]: time="2026-03-02T14:29:09.396761493Z" level=info msg="connecting to shim a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37" address="unix:///run/containerd/s/bbe57f0c84aceadcc8448ff4b49180b8d039764b07ea3decf6c59664d6ad9d8e" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:29:09.616918 systemd[1]: Started cri-containerd-a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37.scope - libcontainer container a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37. Mar 2 14:29:09.822157 systemd-resolved[1474]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 14:29:10.113076 containerd[1574]: time="2026-03-02T14:29:10.112637469Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fgxfk,Uid:f81f0355-d756-4ec2-b9d0-9639b176b2ea,Namespace:calico-system,Attempt:0,} returns sandbox id \"a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37\"" Mar 2 14:29:10.617108 systemd-networkd[1473]: cali59f3669a390: Gained IPv6LL Mar 2 14:29:10.706541 systemd-networkd[1473]: cali8718d97cd70: Link UP Mar 2 14:29:10.738635 systemd-networkd[1473]: cali8718d97cd70: Gained carrier Mar 2 14:29:10.946312 containerd[1574]: 2026-03-02 14:29:09.666 [INFO][4816] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--66bc5c9577--xrdgw-eth0 coredns-66bc5c9577- kube-system 71148f0a-7223-4509-99b6-37365349cb1b 1048 0 2026-03-02 14:27:21 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-66bc5c9577-xrdgw eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali8718d97cd70 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Namespace="kube-system" Pod="coredns-66bc5c9577-xrdgw" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xrdgw-" Mar 2 14:29:10.946312 containerd[1574]: 2026-03-02 14:29:09.672 [INFO][4816] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Namespace="kube-system" Pod="coredns-66bc5c9577-xrdgw" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" Mar 2 14:29:10.946312 containerd[1574]: 2026-03-02 14:29:10.120 [INFO][4897] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" HandleID="k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Workload="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.168 [INFO][4897] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" HandleID="k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Workload="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0006dc950), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-66bc5c9577-xrdgw", "timestamp":"2026-03-02 14:29:10.120047153 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0006e6420)} Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.168 [INFO][4897] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.168 [INFO][4897] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.168 [INFO][4897] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.226 [INFO][4897] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" host="localhost" Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.343 [INFO][4897] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.391 [INFO][4897] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.424 [INFO][4897] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.449 [INFO][4897] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:10.947005 containerd[1574]: 2026-03-02 14:29:10.450 [INFO][4897] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" host="localhost" Mar 2 14:29:10.947374 containerd[1574]: 2026-03-02 14:29:10.467 [INFO][4897] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457 Mar 2 14:29:10.947374 containerd[1574]: 2026-03-02 14:29:10.502 [INFO][4897] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" host="localhost" Mar 2 14:29:10.947374 containerd[1574]: 2026-03-02 14:29:10.610 [INFO][4897] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" host="localhost" Mar 2 14:29:10.947374 containerd[1574]: 2026-03-02 14:29:10.611 [INFO][4897] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" host="localhost" Mar 2 14:29:10.947374 containerd[1574]: 2026-03-02 14:29:10.625 [INFO][4897] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 14:29:10.947374 containerd[1574]: 2026-03-02 14:29:10.625 [INFO][4897] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" HandleID="k8s-pod-network.dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Workload="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" Mar 2 14:29:10.947622 containerd[1574]: 2026-03-02 14:29:10.671 [INFO][4816] cni-plugin/k8s.go 418: Populated endpoint ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Namespace="kube-system" Pod="coredns-66bc5c9577-xrdgw" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--xrdgw-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"71148f0a-7223-4509-99b6-37365349cb1b", ResourceVersion:"1048", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-66bc5c9577-xrdgw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali8718d97cd70", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:10.947622 containerd[1574]: 2026-03-02 14:29:10.672 [INFO][4816] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Namespace="kube-system" Pod="coredns-66bc5c9577-xrdgw" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" Mar 2 14:29:10.947622 containerd[1574]: 2026-03-02 14:29:10.672 [INFO][4816] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8718d97cd70 ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Namespace="kube-system" Pod="coredns-66bc5c9577-xrdgw" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" Mar 2 14:29:10.947622 containerd[1574]: 2026-03-02 14:29:10.756 [INFO][4816] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Namespace="kube-system" Pod="coredns-66bc5c9577-xrdgw" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" Mar 2 14:29:10.947622 containerd[1574]: 2026-03-02 14:29:10.767 [INFO][4816] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Namespace="kube-system" Pod="coredns-66bc5c9577-xrdgw" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--xrdgw-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"71148f0a-7223-4509-99b6-37365349cb1b", ResourceVersion:"1048", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457", Pod:"coredns-66bc5c9577-xrdgw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali8718d97cd70", MAC:"6a:86:02:bd:93:64", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:10.947622 containerd[1574]: 2026-03-02 14:29:10.925 [INFO][4816] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" Namespace="kube-system" Pod="coredns-66bc5c9577-xrdgw" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xrdgw-eth0" Mar 2 14:29:11.238019 containerd[1574]: time="2026-03-02T14:29:11.237772509Z" level=info msg="connecting to shim dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457" address="unix:///run/containerd/s/e646ce3ef977af0fabcdc9c4f0a9e0892b292d3bc18cfe1000527c637614e8a4" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:29:11.264161 systemd-networkd[1473]: cali21a2f4fc42b: Link UP Mar 2 14:29:11.270118 systemd-networkd[1473]: cali21a2f4fc42b: Gained carrier Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:09.657 [INFO][4793] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0 calico-apiserver-68647fc8d9- calico-system 0dd516cb-8822-4937-b80a-6a7395992c08 1053 0 2026-03-02 14:27:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:68647fc8d9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-68647fc8d9-mvfxc eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali21a2f4fc42b [] [] }} ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-mvfxc" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:09.657 [INFO][4793] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-mvfxc" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:10.203 [INFO][4904] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" HandleID="k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Workload="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:10.293 [INFO][4904] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" HandleID="k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Workload="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000471020), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-apiserver-68647fc8d9-mvfxc", "timestamp":"2026-03-02 14:29:10.20326009 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0001ee6e0)} Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:10.294 [INFO][4904] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:10.630 [INFO][4904] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:10.630 [INFO][4904] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:10.727 [INFO][4904] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:10.795 [INFO][4904] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:10.923 [INFO][4904] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.006 [INFO][4904] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.035 [INFO][4904] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.035 [INFO][4904] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.060 [INFO][4904] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.150 [INFO][4904] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.211 [INFO][4904] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.224 [INFO][4904] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" host="localhost" Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.224 [INFO][4904] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 14:29:11.474089 containerd[1574]: 2026-03-02 14:29:11.225 [INFO][4904] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" HandleID="k8s-pod-network.524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Workload="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" Mar 2 14:29:11.475788 containerd[1574]: 2026-03-02 14:29:11.251 [INFO][4793] cni-plugin/k8s.go 418: Populated endpoint ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-mvfxc" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0", GenerateName:"calico-apiserver-68647fc8d9-", Namespace:"calico-system", SelfLink:"", UID:"0dd516cb-8822-4937-b80a-6a7395992c08", ResourceVersion:"1053", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"68647fc8d9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-68647fc8d9-mvfxc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali21a2f4fc42b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:11.475788 containerd[1574]: 2026-03-02 14:29:11.252 [INFO][4793] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-mvfxc" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" Mar 2 14:29:11.475788 containerd[1574]: 2026-03-02 14:29:11.252 [INFO][4793] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali21a2f4fc42b ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-mvfxc" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" Mar 2 14:29:11.475788 containerd[1574]: 2026-03-02 14:29:11.269 [INFO][4793] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-mvfxc" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" Mar 2 14:29:11.475788 containerd[1574]: 2026-03-02 14:29:11.274 [INFO][4793] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-mvfxc" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0", GenerateName:"calico-apiserver-68647fc8d9-", Namespace:"calico-system", SelfLink:"", UID:"0dd516cb-8822-4937-b80a-6a7395992c08", ResourceVersion:"1053", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"68647fc8d9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d", Pod:"calico-apiserver-68647fc8d9-mvfxc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali21a2f4fc42b", MAC:"12:42:6e:61:f4:61", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:11.475788 containerd[1574]: 2026-03-02 14:29:11.465 [INFO][4793] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-mvfxc" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--mvfxc-eth0" Mar 2 14:29:11.508640 systemd[1]: Started cri-containerd-dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457.scope - libcontainer container dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457. Mar 2 14:29:11.609659 systemd-resolved[1474]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 14:29:11.778970 containerd[1574]: time="2026-03-02T14:29:11.778723704Z" level=info msg="connecting to shim 524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d" address="unix:///run/containerd/s/0cddbe0fbaa532cc078bc7f097203373ca16caea685853f066084ac656e28eae" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:29:11.833939 systemd-networkd[1473]: cali22e70fb4cdf: Link UP Mar 2 14:29:11.835331 systemd-networkd[1473]: cali22e70fb4cdf: Gained carrier Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:09.967 [INFO][4841] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--66bc5c9577--xlqmd-eth0 coredns-66bc5c9577- kube-system bc8db32a-755c-44fc-abe9-e927df38b7b5 1051 0 2026-03-02 14:27:21 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-66bc5c9577-xlqmd eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali22e70fb4cdf [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Namespace="kube-system" Pod="coredns-66bc5c9577-xlqmd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xlqmd-" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:09.971 [INFO][4841] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Namespace="kube-system" Pod="coredns-66bc5c9577-xlqmd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:10.310 [INFO][4926] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" HandleID="k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Workload="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:10.345 [INFO][4926] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" HandleID="k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Workload="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000595a00), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-66bc5c9577-xlqmd", "timestamp":"2026-03-02 14:29:10.310369784 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00045c420)} Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:10.345 [INFO][4926] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.224 [INFO][4926] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.224 [INFO][4926] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.365 [INFO][4926] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.453 [INFO][4926] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.538 [INFO][4926] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.561 [INFO][4926] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.576 [INFO][4926] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.576 [INFO][4926] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.599 [INFO][4926] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.679 [INFO][4926] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.763 [INFO][4926] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.763 [INFO][4926] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" host="localhost" Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.763 [INFO][4926] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 14:29:12.023312 containerd[1574]: 2026-03-02 14:29:11.763 [INFO][4926] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" HandleID="k8s-pod-network.2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Workload="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" Mar 2 14:29:12.024739 containerd[1574]: 2026-03-02 14:29:11.781 [INFO][4841] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Namespace="kube-system" Pod="coredns-66bc5c9577-xlqmd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--xlqmd-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"bc8db32a-755c-44fc-abe9-e927df38b7b5", ResourceVersion:"1051", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-66bc5c9577-xlqmd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali22e70fb4cdf", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:12.024739 containerd[1574]: 2026-03-02 14:29:11.781 [INFO][4841] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Namespace="kube-system" Pod="coredns-66bc5c9577-xlqmd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" Mar 2 14:29:12.024739 containerd[1574]: 2026-03-02 14:29:11.781 [INFO][4841] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali22e70fb4cdf ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Namespace="kube-system" Pod="coredns-66bc5c9577-xlqmd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" Mar 2 14:29:12.024739 containerd[1574]: 2026-03-02 14:29:11.834 [INFO][4841] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Namespace="kube-system" Pod="coredns-66bc5c9577-xlqmd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" Mar 2 14:29:12.024739 containerd[1574]: 2026-03-02 14:29:11.836 [INFO][4841] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Namespace="kube-system" Pod="coredns-66bc5c9577-xlqmd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--xlqmd-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"bc8db32a-755c-44fc-abe9-e927df38b7b5", ResourceVersion:"1051", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d", Pod:"coredns-66bc5c9577-xlqmd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali22e70fb4cdf", MAC:"1e:c8:32:5f:01:15", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:12.024739 containerd[1574]: 2026-03-02 14:29:11.935 [INFO][4841] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" Namespace="kube-system" Pod="coredns-66bc5c9577-xlqmd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--xlqmd-eth0" Mar 2 14:29:12.124167 containerd[1574]: time="2026-03-02T14:29:12.121542745Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-xrdgw,Uid:71148f0a-7223-4509-99b6-37365349cb1b,Namespace:kube-system,Attempt:0,} returns sandbox id \"dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457\"" Mar 2 14:29:12.141331 systemd[1]: Started cri-containerd-524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d.scope - libcontainer container 524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d. Mar 2 14:29:12.170764 kubelet[2823]: E0302 14:29:12.144377 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:12.193519 containerd[1574]: time="2026-03-02T14:29:12.192952535Z" level=info msg="CreateContainer within sandbox \"dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 14:29:12.306185 containerd[1574]: time="2026-03-02T14:29:12.306011185Z" level=info msg="connecting to shim 2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d" address="unix:///run/containerd/s/6ec8fcad601181a8228bdbb8191286962b9811205b584d356bb8bcdbb276eaf1" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:29:12.309199 systemd-networkd[1473]: calia170c22ddb3: Link UP Mar 2 14:29:12.314006 systemd-networkd[1473]: calia170c22ddb3: Gained carrier Mar 2 14:29:12.335542 systemd-networkd[1473]: cali8718d97cd70: Gained IPv6LL Mar 2 14:29:12.390932 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1973045033.mount: Deactivated successfully. Mar 2 14:29:12.397378 containerd[1574]: time="2026-03-02T14:29:12.395762368Z" level=info msg="Container b55d149d290130b0f69110481be56491ac40cd39def83195136842e7a50cf0b6: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:12.405185 systemd-resolved[1474]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:09.807 [INFO][4803] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0 calico-apiserver-68647fc8d9- calico-system ceacecf8-f59e-4263-8232-2bc003edccd4 1057 0 2026-03-02 14:27:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:68647fc8d9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-68647fc8d9-p874s eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] calia170c22ddb3 [] [] }} ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-p874s" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--p874s-" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:09.807 [INFO][4803] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-p874s" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:10.227 [INFO][4911] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" HandleID="k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Workload="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:10.376 [INFO][4911] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" HandleID="k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Workload="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000502e70), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-apiserver-68647fc8d9-p874s", "timestamp":"2026-03-02 14:29:10.227718914 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00068e9a0)} Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:10.376 [INFO][4911] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:11.765 [INFO][4911] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:11.765 [INFO][4911] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:11.791 [INFO][4911] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:11.839 [INFO][4911] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:11.928 [INFO][4911] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:11.977 [INFO][4911] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:12.051 [INFO][4911] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:12.052 [INFO][4911] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:12.073 [INFO][4911] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99 Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:12.144 [INFO][4911] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:12.213 [INFO][4911] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:12.213 [INFO][4911] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" host="localhost" Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:12.213 [INFO][4911] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 14:29:12.460106 containerd[1574]: 2026-03-02 14:29:12.213 [INFO][4911] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" HandleID="k8s-pod-network.3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Workload="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" Mar 2 14:29:12.465758 containerd[1574]: 2026-03-02 14:29:12.235 [INFO][4803] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-p874s" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0", GenerateName:"calico-apiserver-68647fc8d9-", Namespace:"calico-system", SelfLink:"", UID:"ceacecf8-f59e-4263-8232-2bc003edccd4", ResourceVersion:"1057", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"68647fc8d9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-68647fc8d9-p874s", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calia170c22ddb3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:12.465758 containerd[1574]: 2026-03-02 14:29:12.235 [INFO][4803] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-p874s" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" Mar 2 14:29:12.465758 containerd[1574]: 2026-03-02 14:29:12.236 [INFO][4803] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia170c22ddb3 ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-p874s" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" Mar 2 14:29:12.465758 containerd[1574]: 2026-03-02 14:29:12.313 [INFO][4803] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-p874s" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" Mar 2 14:29:12.465758 containerd[1574]: 2026-03-02 14:29:12.325 [INFO][4803] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-p874s" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0", GenerateName:"calico-apiserver-68647fc8d9-", Namespace:"calico-system", SelfLink:"", UID:"ceacecf8-f59e-4263-8232-2bc003edccd4", ResourceVersion:"1057", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 14, 27, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"68647fc8d9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99", Pod:"calico-apiserver-68647fc8d9-p874s", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calia170c22ddb3", MAC:"7a:c3:b1:81:ae:d4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 14:29:12.465758 containerd[1574]: 2026-03-02 14:29:12.423 [INFO][4803] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" Namespace="calico-system" Pod="calico-apiserver-68647fc8d9-p874s" WorkloadEndpoint="localhost-k8s-calico--apiserver--68647fc8d9--p874s-eth0" Mar 2 14:29:12.466160 containerd[1574]: time="2026-03-02T14:29:12.465785259Z" level=info msg="CreateContainer within sandbox \"dd65177ed9ff15fca4d81da5d819423f874fc51f53810fbe567be3fa06c4b457\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b55d149d290130b0f69110481be56491ac40cd39def83195136842e7a50cf0b6\"" Mar 2 14:29:12.485720 containerd[1574]: time="2026-03-02T14:29:12.480004167Z" level=info msg="StartContainer for \"b55d149d290130b0f69110481be56491ac40cd39def83195136842e7a50cf0b6\"" Mar 2 14:29:12.485720 containerd[1574]: time="2026-03-02T14:29:12.481250271Z" level=info msg="connecting to shim b55d149d290130b0f69110481be56491ac40cd39def83195136842e7a50cf0b6" address="unix:///run/containerd/s/e646ce3ef977af0fabcdc9c4f0a9e0892b292d3bc18cfe1000527c637614e8a4" protocol=ttrpc version=3 Mar 2 14:29:12.604233 systemd[1]: Started cri-containerd-2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d.scope - libcontainer container 2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d. Mar 2 14:29:12.623869 containerd[1574]: time="2026-03-02T14:29:12.623580369Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68647fc8d9-mvfxc,Uid:0dd516cb-8822-4937-b80a-6a7395992c08,Namespace:calico-system,Attempt:0,} returns sandbox id \"524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d\"" Mar 2 14:29:12.666550 containerd[1574]: time="2026-03-02T14:29:12.660400432Z" level=info msg="connecting to shim 3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99" address="unix:///run/containerd/s/ef60fec5ffbe10139033202b15ce65140ff9565c037675a919b226e6e6eaa53c" namespace=k8s.io protocol=ttrpc version=3 Mar 2 14:29:12.739691 systemd[1]: Started cri-containerd-b55d149d290130b0f69110481be56491ac40cd39def83195136842e7a50cf0b6.scope - libcontainer container b55d149d290130b0f69110481be56491ac40cd39def83195136842e7a50cf0b6. Mar 2 14:29:12.797118 systemd-resolved[1474]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 14:29:12.879148 systemd[1]: Started cri-containerd-3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99.scope - libcontainer container 3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99. Mar 2 14:29:13.106186 systemd-networkd[1473]: cali21a2f4fc42b: Gained IPv6LL Mar 2 14:29:13.208079 systemd-resolved[1474]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 14:29:13.317001 containerd[1574]: time="2026-03-02T14:29:13.316947524Z" level=info msg="StartContainer for \"b55d149d290130b0f69110481be56491ac40cd39def83195136842e7a50cf0b6\" returns successfully" Mar 2 14:29:13.331559 containerd[1574]: time="2026-03-02T14:29:13.331249377Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-xlqmd,Uid:bc8db32a-755c-44fc-abe9-e927df38b7b5,Namespace:kube-system,Attempt:0,} returns sandbox id \"2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d\"" Mar 2 14:29:13.335606 kubelet[2823]: E0302 14:29:13.335247 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:13.439340 containerd[1574]: time="2026-03-02T14:29:13.398358817Z" level=info msg="CreateContainer within sandbox \"2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 14:29:13.496612 systemd-networkd[1473]: cali22e70fb4cdf: Gained IPv6LL Mar 2 14:29:13.540812 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount696750928.mount: Deactivated successfully. Mar 2 14:29:13.548812 kubelet[2823]: E0302 14:29:13.548681 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:13.594290 containerd[1574]: time="2026-03-02T14:29:13.590176041Z" level=info msg="Container d368a8e1a5ddb5b9b391d12a042127969183cdb5fb1093c9b6ec6c17cf78e8dd: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:13.678104 systemd-networkd[1473]: calia170c22ddb3: Gained IPv6LL Mar 2 14:29:13.727585 containerd[1574]: time="2026-03-02T14:29:13.726652502Z" level=info msg="CreateContainer within sandbox \"2b761f5fde1589e31db226e80627e8c91bcb1d900265384357fd9d214db58e3d\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d368a8e1a5ddb5b9b391d12a042127969183cdb5fb1093c9b6ec6c17cf78e8dd\"" Mar 2 14:29:13.734567 containerd[1574]: time="2026-03-02T14:29:13.734498480Z" level=info msg="StartContainer for \"d368a8e1a5ddb5b9b391d12a042127969183cdb5fb1093c9b6ec6c17cf78e8dd\"" Mar 2 14:29:13.736154 containerd[1574]: time="2026-03-02T14:29:13.735811398Z" level=info msg="connecting to shim d368a8e1a5ddb5b9b391d12a042127969183cdb5fb1093c9b6ec6c17cf78e8dd" address="unix:///run/containerd/s/6ec8fcad601181a8228bdbb8191286962b9811205b584d356bb8bcdbb276eaf1" protocol=ttrpc version=3 Mar 2 14:29:13.820338 kubelet[2823]: I0302 14:29:13.820275 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-xrdgw" podStartSLOduration=112.820258772 podStartE2EDuration="1m52.820258772s" podCreationTimestamp="2026-03-02 14:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 14:29:13.818945174 +0000 UTC m=+117.252613216" watchObservedRunningTime="2026-03-02 14:29:13.820258772 +0000 UTC m=+117.253926804" Mar 2 14:29:13.905359 systemd[1]: Started sshd@9-10.0.0.9:22-10.0.0.1:36012.service - OpenSSH per-connection server daemon (10.0.0.1:36012). Mar 2 14:29:13.981322 containerd[1574]: time="2026-03-02T14:29:13.978425701Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68647fc8d9-p874s,Uid:ceacecf8-f59e-4263-8232-2bc003edccd4,Namespace:calico-system,Attempt:0,} returns sandbox id \"3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99\"" Mar 2 14:29:13.978798 systemd[1]: Started cri-containerd-d368a8e1a5ddb5b9b391d12a042127969183cdb5fb1093c9b6ec6c17cf78e8dd.scope - libcontainer container d368a8e1a5ddb5b9b391d12a042127969183cdb5fb1093c9b6ec6c17cf78e8dd. Mar 2 14:29:14.359722 containerd[1574]: time="2026-03-02T14:29:14.353611507Z" level=info msg="StartContainer for \"d368a8e1a5ddb5b9b391d12a042127969183cdb5fb1093c9b6ec6c17cf78e8dd\" returns successfully" Mar 2 14:29:14.410723 sshd[5257]: Accepted publickey for core from 10.0.0.1 port 36012 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:29:14.431956 sshd-session[5257]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:29:14.464691 systemd-logind[1554]: New session 10 of user core. Mar 2 14:29:14.477783 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 2 14:29:14.720338 kubelet[2823]: E0302 14:29:14.712763 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:14.793248 kubelet[2823]: E0302 14:29:14.793139 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:15.578036 sshd[5290]: Connection closed by 10.0.0.1 port 36012 Mar 2 14:29:15.580428 sshd-session[5257]: pam_unix(sshd:session): session closed for user core Mar 2 14:29:15.604818 systemd[1]: sshd@9-10.0.0.9:22-10.0.0.1:36012.service: Deactivated successfully. Mar 2 14:29:15.617679 systemd[1]: session-10.scope: Deactivated successfully. Mar 2 14:29:15.639884 systemd-logind[1554]: Session 10 logged out. Waiting for processes to exit. Mar 2 14:29:15.647521 systemd-logind[1554]: Removed session 10. Mar 2 14:29:15.781397 kubelet[2823]: E0302 14:29:15.780699 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:15.783093 kubelet[2823]: E0302 14:29:15.783015 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:15.962742 kubelet[2823]: I0302 14:29:15.956784 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-xlqmd" podStartSLOduration=114.956761445 podStartE2EDuration="1m54.956761445s" podCreationTimestamp="2026-03-02 14:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 14:29:14.953159318 +0000 UTC m=+118.386827331" watchObservedRunningTime="2026-03-02 14:29:15.956761445 +0000 UTC m=+119.390429448" Mar 2 14:29:16.798973 kubelet[2823]: E0302 14:29:16.797708 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:16.802518 kubelet[2823]: E0302 14:29:16.801322 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:19.936494 containerd[1574]: time="2026-03-02T14:29:19.935160746Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:19.962145 containerd[1574]: time="2026-03-02T14:29:19.959513480Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.3: active requests=0, bytes read=52396348" Mar 2 14:29:19.968419 containerd[1574]: time="2026-03-02T14:29:19.968325036Z" level=info msg="ImageCreate event name:\"sha256:95bc8e4bc61e762d7451304ff00b4ebc2aed857d8698340cb94b885328290dfe\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:19.981113 containerd[1574]: time="2026-03-02T14:29:19.980910964Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:081fd6c3de7754ba9892532b2c7c6cae9ba7bd1cca4c42e4590ee8d0f5a5696b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:19.983951 containerd[1574]: time="2026-03-02T14:29:19.981686159Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\" with image id \"sha256:95bc8e4bc61e762d7451304ff00b4ebc2aed857d8698340cb94b885328290dfe\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:081fd6c3de7754ba9892532b2c7c6cae9ba7bd1cca4c42e4590ee8d0f5a5696b\", size \"53952361\" in 13.140367503s" Mar 2 14:29:19.983951 containerd[1574]: time="2026-03-02T14:29:19.981719220Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\" returns image reference \"sha256:95bc8e4bc61e762d7451304ff00b4ebc2aed857d8698340cb94b885328290dfe\"" Mar 2 14:29:20.025998 containerd[1574]: time="2026-03-02T14:29:20.022530162Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.3\"" Mar 2 14:29:20.070942 containerd[1574]: time="2026-03-02T14:29:20.070054204Z" level=info msg="CreateContainer within sandbox \"5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Mar 2 14:29:20.160208 containerd[1574]: time="2026-03-02T14:29:20.160124600Z" level=info msg="Container bc1471524e083c7d3bd00c77d1054075fbe0202b043793f42345e2fbb3f355da: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:20.218935 containerd[1574]: time="2026-03-02T14:29:20.214490582Z" level=info msg="CreateContainer within sandbox \"5b9cafea39589a6c692c98cf4d6d90888e495d40a7cb54aa0c1c1fb0274613bc\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"bc1471524e083c7d3bd00c77d1054075fbe0202b043793f42345e2fbb3f355da\"" Mar 2 14:29:20.219277 containerd[1574]: time="2026-03-02T14:29:20.219245817Z" level=info msg="StartContainer for \"bc1471524e083c7d3bd00c77d1054075fbe0202b043793f42345e2fbb3f355da\"" Mar 2 14:29:20.228277 containerd[1574]: time="2026-03-02T14:29:20.228241480Z" level=info msg="connecting to shim bc1471524e083c7d3bd00c77d1054075fbe0202b043793f42345e2fbb3f355da" address="unix:///run/containerd/s/1a6e4873526a1d1e5d4fc85305ba66537be9e818ce17d96defc3c65d056b3dc2" protocol=ttrpc version=3 Mar 2 14:29:20.392946 systemd[1]: Started cri-containerd-bc1471524e083c7d3bd00c77d1054075fbe0202b043793f42345e2fbb3f355da.scope - libcontainer container bc1471524e083c7d3bd00c77d1054075fbe0202b043793f42345e2fbb3f355da. Mar 2 14:29:20.638589 systemd[1]: Started sshd@10-10.0.0.9:22-10.0.0.1:41076.service - OpenSSH per-connection server daemon (10.0.0.1:41076). Mar 2 14:29:20.907968 containerd[1574]: time="2026-03-02T14:29:20.904313148Z" level=info msg="StartContainer for \"bc1471524e083c7d3bd00c77d1054075fbe0202b043793f42345e2fbb3f355da\" returns successfully" Mar 2 14:29:21.145046 sshd[5359]: Accepted publickey for core from 10.0.0.1 port 41076 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:29:21.161589 sshd-session[5359]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:29:21.201728 systemd-logind[1554]: New session 11 of user core. Mar 2 14:29:21.242363 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 2 14:29:22.113152 kubelet[2823]: E0302 14:29:22.112925 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:22.136807 sshd[5382]: Connection closed by 10.0.0.1 port 41076 Mar 2 14:29:22.141893 sshd-session[5359]: pam_unix(sshd:session): session closed for user core Mar 2 14:29:22.150385 systemd[1]: sshd@10-10.0.0.9:22-10.0.0.1:41076.service: Deactivated successfully. Mar 2 14:29:22.156305 systemd[1]: session-11.scope: Deactivated successfully. Mar 2 14:29:22.160758 systemd-logind[1554]: Session 11 logged out. Waiting for processes to exit. Mar 2 14:29:22.166803 systemd-logind[1554]: Removed session 11. Mar 2 14:29:23.638658 kubelet[2823]: I0302 14:29:23.638300 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-6d7c47ccd9-vmfz4" podStartSLOduration=80.485122018 podStartE2EDuration="1m33.638281457s" podCreationTimestamp="2026-03-02 14:27:50 +0000 UTC" firstStartedPulling="2026-03-02 14:29:06.839270737 +0000 UTC m=+110.272938741" lastFinishedPulling="2026-03-02 14:29:19.992430178 +0000 UTC m=+123.426098180" observedRunningTime="2026-03-02 14:29:22.090985362 +0000 UTC m=+125.524653395" watchObservedRunningTime="2026-03-02 14:29:23.638281457 +0000 UTC m=+127.071949460" Mar 2 14:29:27.175665 systemd[1]: Started sshd@11-10.0.0.9:22-10.0.0.1:41092.service - OpenSSH per-connection server daemon (10.0.0.1:41092). Mar 2 14:29:27.411306 sshd[5481]: Accepted publickey for core from 10.0.0.1 port 41092 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:29:27.413498 sshd-session[5481]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:29:27.428614 systemd-logind[1554]: New session 12 of user core. Mar 2 14:29:27.450431 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 2 14:29:28.374961 sshd[5484]: Connection closed by 10.0.0.1 port 41092 Mar 2 14:29:28.371684 sshd-session[5481]: pam_unix(sshd:session): session closed for user core Mar 2 14:29:28.398841 systemd[1]: sshd@11-10.0.0.9:22-10.0.0.1:41092.service: Deactivated successfully. Mar 2 14:29:28.422991 systemd[1]: session-12.scope: Deactivated successfully. Mar 2 14:29:28.437306 systemd-logind[1554]: Session 12 logged out. Waiting for processes to exit. Mar 2 14:29:28.444670 systemd-logind[1554]: Removed session 12. Mar 2 14:29:28.625107 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3348406991.mount: Deactivated successfully. Mar 2 14:29:32.032955 containerd[1574]: time="2026-03-02T14:29:32.031639894Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:32.045820 containerd[1574]: time="2026-03-02T14:29:32.045525676Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.3: active requests=0, bytes read=55607954" Mar 2 14:29:32.063366 containerd[1574]: time="2026-03-02T14:29:32.060739605Z" level=info msg="ImageCreate event name:\"sha256:6eaae458d5f115c04bbd6cd0facdbc393958d24af9934b90825fea68960a2f1a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:32.085413 containerd[1574]: time="2026-03-02T14:29:32.077831665Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:e85ffa1d9468908b0bd44664de0d023da6669faefb3e1013b3a15b63dfa1f9a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:32.085413 containerd[1574]: time="2026-03-02T14:29:32.078562588Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.3\" with image id \"sha256:6eaae458d5f115c04bbd6cd0facdbc393958d24af9934b90825fea68960a2f1a\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:e85ffa1d9468908b0bd44664de0d023da6669faefb3e1013b3a15b63dfa1f9a9\", size \"55607800\" in 12.055754239s" Mar 2 14:29:32.085413 containerd[1574]: time="2026-03-02T14:29:32.082037678Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.3\" returns image reference \"sha256:6eaae458d5f115c04bbd6cd0facdbc393958d24af9934b90825fea68960a2f1a\"" Mar 2 14:29:32.126152 containerd[1574]: time="2026-03-02T14:29:32.123143927Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.3\"" Mar 2 14:29:32.150185 containerd[1574]: time="2026-03-02T14:29:32.148997976Z" level=info msg="CreateContainer within sandbox \"a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Mar 2 14:29:32.451389 containerd[1574]: time="2026-03-02T14:29:32.434603901Z" level=info msg="Container 5a1d284f4a5524dcedada775a65872145272f5df2774315686783ea5497b0dbd: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:32.662963 containerd[1574]: time="2026-03-02T14:29:32.660007529Z" level=info msg="CreateContainer within sandbox \"a73b228449b1a6ca8b9ec330d157bd18111c4b0d5e9f36da670a207239e208ec\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"5a1d284f4a5524dcedada775a65872145272f5df2774315686783ea5497b0dbd\"" Mar 2 14:29:32.666994 containerd[1574]: time="2026-03-02T14:29:32.664917985Z" level=info msg="StartContainer for \"5a1d284f4a5524dcedada775a65872145272f5df2774315686783ea5497b0dbd\"" Mar 2 14:29:32.669426 containerd[1574]: time="2026-03-02T14:29:32.669348106Z" level=info msg="connecting to shim 5a1d284f4a5524dcedada775a65872145272f5df2774315686783ea5497b0dbd" address="unix:///run/containerd/s/c22c5d2253d8d5372b7beb2dba1a21e665e1914cc786b8cee5ae88a5e7554d57" protocol=ttrpc version=3 Mar 2 14:29:32.931625 systemd[1]: Started cri-containerd-5a1d284f4a5524dcedada775a65872145272f5df2774315686783ea5497b0dbd.scope - libcontainer container 5a1d284f4a5524dcedada775a65872145272f5df2774315686783ea5497b0dbd. Mar 2 14:29:33.166811 containerd[1574]: time="2026-03-02T14:29:33.159533917Z" level=info msg="StartContainer for \"5a1d284f4a5524dcedada775a65872145272f5df2774315686783ea5497b0dbd\" returns successfully" Mar 2 14:29:33.420426 systemd[1]: Started sshd@12-10.0.0.9:22-10.0.0.1:35694.service - OpenSSH per-connection server daemon (10.0.0.1:35694). Mar 2 14:29:33.755248 sshd[5551]: Accepted publickey for core from 10.0.0.1 port 35694 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:29:33.758723 sshd-session[5551]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:29:33.777293 systemd-logind[1554]: New session 13 of user core. Mar 2 14:29:33.805376 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 2 14:29:34.364492 containerd[1574]: time="2026-03-02T14:29:34.364395778Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:34.372727 containerd[1574]: time="2026-03-02T14:29:34.372631282Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.3: active requests=0, bytes read=8793087" Mar 2 14:29:34.379346 containerd[1574]: time="2026-03-02T14:29:34.378576959Z" level=info msg="ImageCreate event name:\"sha256:6f60b868a297033aea2daba09eb6f77fb2390c659bbc8dfaaac24f32f5b84e27\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:34.397409 containerd[1574]: time="2026-03-02T14:29:34.397357558Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:3d04cd6265f850f0420b413351275ebfd244991b1b9e69c64efe8b4eff45b53f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:34.401505 containerd[1574]: time="2026-03-02T14:29:34.399508837Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.3\" with image id \"sha256:6f60b868a297033aea2daba09eb6f77fb2390c659bbc8dfaaac24f32f5b84e27\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:3d04cd6265f850f0420b413351275ebfd244991b1b9e69c64efe8b4eff45b53f\", size \"10349132\" in 2.276237072s" Mar 2 14:29:34.401505 containerd[1574]: time="2026-03-02T14:29:34.399546498Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.3\" returns image reference \"sha256:6f60b868a297033aea2daba09eb6f77fb2390c659bbc8dfaaac24f32f5b84e27\"" Mar 2 14:29:34.409283 containerd[1574]: time="2026-03-02T14:29:34.409249181Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\"" Mar 2 14:29:34.433784 containerd[1574]: time="2026-03-02T14:29:34.433740669Z" level=info msg="CreateContainer within sandbox \"a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 2 14:29:34.507834 sshd[5560]: Connection closed by 10.0.0.1 port 35694 Mar 2 14:29:34.509188 sshd-session[5551]: pam_unix(sshd:session): session closed for user core Mar 2 14:29:34.521945 containerd[1574]: time="2026-03-02T14:29:34.521720325Z" level=info msg="Container 57a6bcff2ad36dcd6613cbba81c4072676aa45bd519e2d48e215cf01c88344c3: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:34.543546 systemd[1]: sshd@12-10.0.0.9:22-10.0.0.1:35694.service: Deactivated successfully. Mar 2 14:29:34.547321 systemd[1]: session-13.scope: Deactivated successfully. Mar 2 14:29:34.549712 systemd-logind[1554]: Session 13 logged out. Waiting for processes to exit. Mar 2 14:29:34.556139 systemd-logind[1554]: Removed session 13. Mar 2 14:29:34.563766 containerd[1574]: time="2026-03-02T14:29:34.563111886Z" level=info msg="CreateContainer within sandbox \"a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"57a6bcff2ad36dcd6613cbba81c4072676aa45bd519e2d48e215cf01c88344c3\"" Mar 2 14:29:34.568494 containerd[1574]: time="2026-03-02T14:29:34.566692982Z" level=info msg="StartContainer for \"57a6bcff2ad36dcd6613cbba81c4072676aa45bd519e2d48e215cf01c88344c3\"" Mar 2 14:29:34.570849 containerd[1574]: time="2026-03-02T14:29:34.570189805Z" level=info msg="connecting to shim 57a6bcff2ad36dcd6613cbba81c4072676aa45bd519e2d48e215cf01c88344c3" address="unix:///run/containerd/s/bbe57f0c84aceadcc8448ff4b49180b8d039764b07ea3decf6c59664d6ad9d8e" protocol=ttrpc version=3 Mar 2 14:29:34.658937 systemd[1]: Started cri-containerd-57a6bcff2ad36dcd6613cbba81c4072676aa45bd519e2d48e215cf01c88344c3.scope - libcontainer container 57a6bcff2ad36dcd6613cbba81c4072676aa45bd519e2d48e215cf01c88344c3. Mar 2 14:29:35.023114 containerd[1574]: time="2026-03-02T14:29:35.019383245Z" level=info msg="StartContainer for \"57a6bcff2ad36dcd6613cbba81c4072676aa45bd519e2d48e215cf01c88344c3\" returns successfully" Mar 2 14:29:39.557987 systemd[1]: Started sshd@13-10.0.0.9:22-10.0.0.1:35706.service - OpenSSH per-connection server daemon (10.0.0.1:35706). Mar 2 14:29:39.822255 sshd[5660]: Accepted publickey for core from 10.0.0.1 port 35706 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:29:39.833217 sshd-session[5660]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:29:39.852640 systemd-logind[1554]: New session 14 of user core. Mar 2 14:29:39.866801 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 2 14:29:40.704514 sshd[5663]: Connection closed by 10.0.0.1 port 35706 Mar 2 14:29:40.709761 sshd-session[5660]: pam_unix(sshd:session): session closed for user core Mar 2 14:29:40.720807 systemd[1]: sshd@13-10.0.0.9:22-10.0.0.1:35706.service: Deactivated successfully. Mar 2 14:29:40.728075 systemd[1]: session-14.scope: Deactivated successfully. Mar 2 14:29:40.738215 systemd-logind[1554]: Session 14 logged out. Waiting for processes to exit. Mar 2 14:29:40.740732 systemd-logind[1554]: Removed session 14. Mar 2 14:29:43.102515 kubelet[2823]: E0302 14:29:43.101695 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:43.895228 containerd[1574]: time="2026-03-02T14:29:43.895045930Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.3: active requests=0, bytes read=48403149" Mar 2 14:29:43.895228 containerd[1574]: time="2026-03-02T14:29:43.895220785Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:43.915479 containerd[1574]: time="2026-03-02T14:29:43.908031697Z" level=info msg="ImageCreate event name:\"sha256:ac46eecb3d7f840a860cf32547a175e8efb0ec76cc6ff942e75d49177b70c694\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:43.933485 containerd[1574]: time="2026-03-02T14:29:43.927514157Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" with image id \"sha256:ac46eecb3d7f840a860cf32547a175e8efb0ec76cc6ff942e75d49177b70c694\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:c2def03be7412561bd678df17fcf2467cac990dbb42278dcfe193aa5a43128d4\", size \"49959210\" in 9.51809534s" Mar 2 14:29:43.933485 containerd[1574]: time="2026-03-02T14:29:43.927590429Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" returns image reference \"sha256:ac46eecb3d7f840a860cf32547a175e8efb0ec76cc6ff942e75d49177b70c694\"" Mar 2 14:29:43.935854 containerd[1574]: time="2026-03-02T14:29:43.935820545Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:c2def03be7412561bd678df17fcf2467cac990dbb42278dcfe193aa5a43128d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:43.949817 containerd[1574]: time="2026-03-02T14:29:43.949684480Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\"" Mar 2 14:29:43.971917 containerd[1574]: time="2026-03-02T14:29:43.970209756Z" level=info msg="CreateContainer within sandbox \"524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 2 14:29:44.035716 containerd[1574]: time="2026-03-02T14:29:44.033126163Z" level=info msg="Container 2545f98431d457400c7daedde5d6683cc2e07d85216da92ad4b04fd8342250ed: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:44.035423 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1673087116.mount: Deactivated successfully. Mar 2 14:29:44.120954 containerd[1574]: time="2026-03-02T14:29:44.120827134Z" level=info msg="CreateContainer within sandbox \"524fd8734b945be2259aeecc0e458581f5638627cc88fd123775f2e3754c8a6d\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"2545f98431d457400c7daedde5d6683cc2e07d85216da92ad4b04fd8342250ed\"" Mar 2 14:29:44.129566 containerd[1574]: time="2026-03-02T14:29:44.124717529Z" level=info msg="StartContainer for \"2545f98431d457400c7daedde5d6683cc2e07d85216da92ad4b04fd8342250ed\"" Mar 2 14:29:44.129566 containerd[1574]: time="2026-03-02T14:29:44.126306280Z" level=info msg="connecting to shim 2545f98431d457400c7daedde5d6683cc2e07d85216da92ad4b04fd8342250ed" address="unix:///run/containerd/s/0cddbe0fbaa532cc078bc7f097203373ca16caea685853f066084ac656e28eae" protocol=ttrpc version=3 Mar 2 14:29:44.224037 systemd[1]: Started cri-containerd-2545f98431d457400c7daedde5d6683cc2e07d85216da92ad4b04fd8342250ed.scope - libcontainer container 2545f98431d457400c7daedde5d6683cc2e07d85216da92ad4b04fd8342250ed. Mar 2 14:29:44.242923 containerd[1574]: time="2026-03-02T14:29:44.239829187Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:44.248991 containerd[1574]: time="2026-03-02T14:29:44.247626396Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.3: active requests=0, bytes read=77" Mar 2 14:29:44.262701 containerd[1574]: time="2026-03-02T14:29:44.261727781Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" with image id \"sha256:ac46eecb3d7f840a860cf32547a175e8efb0ec76cc6ff942e75d49177b70c694\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:c2def03be7412561bd678df17fcf2467cac990dbb42278dcfe193aa5a43128d4\", size \"49959210\" in 309.284406ms" Mar 2 14:29:44.262701 containerd[1574]: time="2026-03-02T14:29:44.262628750Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" returns image reference \"sha256:ac46eecb3d7f840a860cf32547a175e8efb0ec76cc6ff942e75d49177b70c694\"" Mar 2 14:29:44.279832 containerd[1574]: time="2026-03-02T14:29:44.279741764Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\"" Mar 2 14:29:44.304914 containerd[1574]: time="2026-03-02T14:29:44.304832602Z" level=info msg="CreateContainer within sandbox \"3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 2 14:29:44.373970 containerd[1574]: time="2026-03-02T14:29:44.370604742Z" level=info msg="Container f8cbd0b5fe16f351cae2b5b578903aa042633a81d3b7d279bf408d1967560b4f: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:44.399500 containerd[1574]: time="2026-03-02T14:29:44.396985928Z" level=info msg="CreateContainer within sandbox \"3f3abf8646bcdc29ba7f7669f41483612b3827523a61b6ac4137500eac029b99\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"f8cbd0b5fe16f351cae2b5b578903aa042633a81d3b7d279bf408d1967560b4f\"" Mar 2 14:29:44.408958 containerd[1574]: time="2026-03-02T14:29:44.407332864Z" level=info msg="StartContainer for \"f8cbd0b5fe16f351cae2b5b578903aa042633a81d3b7d279bf408d1967560b4f\"" Mar 2 14:29:44.419670 containerd[1574]: time="2026-03-02T14:29:44.419527745Z" level=info msg="connecting to shim f8cbd0b5fe16f351cae2b5b578903aa042633a81d3b7d279bf408d1967560b4f" address="unix:///run/containerd/s/ef60fec5ffbe10139033202b15ce65140ff9565c037675a919b226e6e6eaa53c" protocol=ttrpc version=3 Mar 2 14:29:44.574201 containerd[1574]: time="2026-03-02T14:29:44.571645880Z" level=info msg="StartContainer for \"2545f98431d457400c7daedde5d6683cc2e07d85216da92ad4b04fd8342250ed\" returns successfully" Mar 2 14:29:44.598575 systemd[1]: Started cri-containerd-f8cbd0b5fe16f351cae2b5b578903aa042633a81d3b7d279bf408d1967560b4f.scope - libcontainer container f8cbd0b5fe16f351cae2b5b578903aa042633a81d3b7d279bf408d1967560b4f. Mar 2 14:29:44.854991 containerd[1574]: time="2026-03-02T14:29:44.854334112Z" level=info msg="StartContainer for \"f8cbd0b5fe16f351cae2b5b578903aa042633a81d3b7d279bf408d1967560b4f\" returns successfully" Mar 2 14:29:45.350056 kubelet[2823]: I0302 14:29:45.349816 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d7f6b6d6-zx66p" podStartSLOduration=97.989777667 podStartE2EDuration="2m1.349796395s" podCreationTimestamp="2026-03-02 14:27:44 +0000 UTC" firstStartedPulling="2026-03-02 14:29:08.746371057 +0000 UTC m=+112.180039060" lastFinishedPulling="2026-03-02 14:29:32.106389785 +0000 UTC m=+135.540057788" observedRunningTime="2026-03-02 14:29:34.191075978 +0000 UTC m=+137.624743981" watchObservedRunningTime="2026-03-02 14:29:45.349796395 +0000 UTC m=+148.783464399" Mar 2 14:29:45.351065 kubelet[2823]: I0302 14:29:45.350151 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-68647fc8d9-p874s" podStartSLOduration=91.068720828 podStartE2EDuration="2m1.350141669s" podCreationTimestamp="2026-03-02 14:27:44 +0000 UTC" firstStartedPulling="2026-03-02 14:29:13.994506472 +0000 UTC m=+117.428174474" lastFinishedPulling="2026-03-02 14:29:44.275927291 +0000 UTC m=+147.709595315" observedRunningTime="2026-03-02 14:29:45.312023198 +0000 UTC m=+148.745691221" watchObservedRunningTime="2026-03-02 14:29:45.350141669 +0000 UTC m=+148.783809672" Mar 2 14:29:45.797560 systemd[1]: Started sshd@14-10.0.0.9:22-10.0.0.1:42068.service - OpenSSH per-connection server daemon (10.0.0.1:42068). Mar 2 14:29:46.457540 sshd[5771]: Accepted publickey for core from 10.0.0.1 port 42068 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:29:46.461549 sshd-session[5771]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:29:46.510126 systemd-logind[1554]: New session 15 of user core. Mar 2 14:29:46.517409 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 2 14:29:47.315978 kubelet[2823]: I0302 14:29:47.315430 2823 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 2 14:29:47.540083 sshd[5778]: Connection closed by 10.0.0.1 port 42068 Mar 2 14:29:47.537524 sshd-session[5771]: pam_unix(sshd:session): session closed for user core Mar 2 14:29:47.568257 systemd[1]: sshd@14-10.0.0.9:22-10.0.0.1:42068.service: Deactivated successfully. Mar 2 14:29:47.582341 systemd[1]: session-15.scope: Deactivated successfully. Mar 2 14:29:47.599560 systemd-logind[1554]: Session 15 logged out. Waiting for processes to exit. Mar 2 14:29:47.622527 systemd-logind[1554]: Removed session 15. Mar 2 14:29:48.107582 kubelet[2823]: E0302 14:29:48.100021 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:29:48.712574 containerd[1574]: time="2026-03-02T14:29:48.711634886Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:48.723200 containerd[1574]: time="2026-03-02T14:29:48.723112462Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3: active requests=0, bytes read=14702266" Mar 2 14:29:48.737274 containerd[1574]: time="2026-03-02T14:29:48.737224414Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\" with image id \"sha256:a06d58cceef55662d827ba735c38dc374717b4fe7115379961a819e177ccc50d\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:2bdced3111efc84af5b77534155b084a55a3f839010807e7e83e75faefc8cf33\", size \"16258263\" in 4.457408252s" Mar 2 14:29:48.737625 containerd[1574]: time="2026-03-02T14:29:48.737504436Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\" returns image reference \"sha256:a06d58cceef55662d827ba735c38dc374717b4fe7115379961a819e177ccc50d\"" Mar 2 14:29:48.763622 containerd[1574]: time="2026-03-02T14:29:48.762053805Z" level=info msg="CreateContainer within sandbox \"a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 2 14:29:48.904562 containerd[1574]: time="2026-03-02T14:29:48.904510820Z" level=info msg="ImageCreate event name:\"sha256:a06d58cceef55662d827ba735c38dc374717b4fe7115379961a819e177ccc50d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:48.910524 containerd[1574]: time="2026-03-02T14:29:48.904827771Z" level=info msg="Container 7ce020e455ee8b435e3210ea38c87899a13e73f3c04cc1642f5eae6ab6bc6c9f: CDI devices from CRI Config.CDIDevices: []" Mar 2 14:29:48.915001 containerd[1574]: time="2026-03-02T14:29:48.914834417Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:2bdced3111efc84af5b77534155b084a55a3f839010807e7e83e75faefc8cf33\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 14:29:48.967673 containerd[1574]: time="2026-03-02T14:29:48.967385665Z" level=info msg="CreateContainer within sandbox \"a152d6e2ed04c2ae996d2396c6694689ea81ca85a54b9c300a2e64f882aa7c37\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"7ce020e455ee8b435e3210ea38c87899a13e73f3c04cc1642f5eae6ab6bc6c9f\"" Mar 2 14:29:48.968777 containerd[1574]: time="2026-03-02T14:29:48.968731612Z" level=info msg="StartContainer for \"7ce020e455ee8b435e3210ea38c87899a13e73f3c04cc1642f5eae6ab6bc6c9f\"" Mar 2 14:29:48.977730 containerd[1574]: time="2026-03-02T14:29:48.977635020Z" level=info msg="connecting to shim 7ce020e455ee8b435e3210ea38c87899a13e73f3c04cc1642f5eae6ab6bc6c9f" address="unix:///run/containerd/s/bbe57f0c84aceadcc8448ff4b49180b8d039764b07ea3decf6c59664d6ad9d8e" protocol=ttrpc version=3 Mar 2 14:29:49.126850 systemd[1]: Started cri-containerd-7ce020e455ee8b435e3210ea38c87899a13e73f3c04cc1642f5eae6ab6bc6c9f.scope - libcontainer container 7ce020e455ee8b435e3210ea38c87899a13e73f3c04cc1642f5eae6ab6bc6c9f. Mar 2 14:29:49.555208 containerd[1574]: time="2026-03-02T14:29:49.555119348Z" level=info msg="StartContainer for \"7ce020e455ee8b435e3210ea38c87899a13e73f3c04cc1642f5eae6ab6bc6c9f\" returns successfully" Mar 2 14:29:49.675495 kubelet[2823]: I0302 14:29:49.675061 2823 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 2 14:29:49.676487 kubelet[2823]: I0302 14:29:49.676313 2823 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 2 14:29:50.550569 kubelet[2823]: I0302 14:29:50.547505 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-68647fc8d9-mvfxc" podStartSLOduration=95.237362198 podStartE2EDuration="2m6.547410835s" podCreationTimestamp="2026-03-02 14:27:44 +0000 UTC" firstStartedPulling="2026-03-02 14:29:12.635682771 +0000 UTC m=+116.069350775" lastFinishedPulling="2026-03-02 14:29:43.945731409 +0000 UTC m=+147.379399412" observedRunningTime="2026-03-02 14:29:45.458801467 +0000 UTC m=+148.892469470" watchObservedRunningTime="2026-03-02 14:29:50.547410835 +0000 UTC m=+153.981078838" Mar 2 14:29:50.557038 kubelet[2823]: I0302 14:29:50.554086 2823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-fgxfk" podStartSLOduration=83.929893539 podStartE2EDuration="2m2.554068921s" podCreationTimestamp="2026-03-02 14:27:48 +0000 UTC" firstStartedPulling="2026-03-02 14:29:10.11675718 +0000 UTC m=+113.550425183" lastFinishedPulling="2026-03-02 14:29:48.740932561 +0000 UTC m=+152.174600565" observedRunningTime="2026-03-02 14:29:50.546223985 +0000 UTC m=+153.979891977" watchObservedRunningTime="2026-03-02 14:29:50.554068921 +0000 UTC m=+153.987736923" Mar 2 14:29:52.593759 systemd[1]: Started sshd@15-10.0.0.9:22-10.0.0.1:36360.service - OpenSSH per-connection server daemon (10.0.0.1:36360). Mar 2 14:29:53.112585 sshd[5834]: Accepted publickey for core from 10.0.0.1 port 36360 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:29:53.127089 sshd-session[5834]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:29:53.188323 systemd-logind[1554]: New session 16 of user core. Mar 2 14:29:53.213715 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 2 14:29:55.147303 sshd[5865]: Connection closed by 10.0.0.1 port 36360 Mar 2 14:29:55.147982 sshd-session[5834]: pam_unix(sshd:session): session closed for user core Mar 2 14:29:55.182391 systemd[1]: sshd@15-10.0.0.9:22-10.0.0.1:36360.service: Deactivated successfully. Mar 2 14:29:55.209159 systemd[1]: session-16.scope: Deactivated successfully. Mar 2 14:29:55.223065 systemd-logind[1554]: Session 16 logged out. Waiting for processes to exit. Mar 2 14:29:55.234738 systemd-logind[1554]: Removed session 16. Mar 2 14:30:00.216779 systemd[1]: Started sshd@16-10.0.0.9:22-10.0.0.1:47382.service - OpenSSH per-connection server daemon (10.0.0.1:47382). Mar 2 14:30:00.508215 sshd[5961]: Accepted publickey for core from 10.0.0.1 port 47382 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:00.516872 sshd-session[5961]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:00.548227 systemd-logind[1554]: New session 17 of user core. Mar 2 14:30:00.604120 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 2 14:30:01.443635 sshd[5964]: Connection closed by 10.0.0.1 port 47382 Mar 2 14:30:01.447997 sshd-session[5961]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:01.473395 systemd[1]: sshd@16-10.0.0.9:22-10.0.0.1:47382.service: Deactivated successfully. Mar 2 14:30:01.482431 systemd[1]: session-17.scope: Deactivated successfully. Mar 2 14:30:01.500736 systemd-logind[1554]: Session 17 logged out. Waiting for processes to exit. Mar 2 14:30:01.513192 systemd-logind[1554]: Removed session 17. Mar 2 14:30:04.117393 kubelet[2823]: E0302 14:30:04.113548 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:30:05.103563 kubelet[2823]: E0302 14:30:05.103494 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:30:06.471885 systemd[1]: Started sshd@17-10.0.0.9:22-10.0.0.1:47398.service - OpenSSH per-connection server daemon (10.0.0.1:47398). Mar 2 14:30:06.668554 sshd[6004]: Accepted publickey for core from 10.0.0.1 port 47398 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:06.673004 sshd-session[6004]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:06.711516 systemd-logind[1554]: New session 18 of user core. Mar 2 14:30:06.741221 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 2 14:30:07.278392 sshd[6007]: Connection closed by 10.0.0.1 port 47398 Mar 2 14:30:07.281117 sshd-session[6004]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:07.306180 systemd[1]: sshd@17-10.0.0.9:22-10.0.0.1:47398.service: Deactivated successfully. Mar 2 14:30:07.315752 systemd[1]: session-18.scope: Deactivated successfully. Mar 2 14:30:07.324964 systemd-logind[1554]: Session 18 logged out. Waiting for processes to exit. Mar 2 14:30:07.344799 systemd-logind[1554]: Removed session 18. Mar 2 14:30:12.321269 systemd[1]: Started sshd@18-10.0.0.9:22-10.0.0.1:39942.service - OpenSSH per-connection server daemon (10.0.0.1:39942). Mar 2 14:30:12.631535 sshd[6023]: Accepted publickey for core from 10.0.0.1 port 39942 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:12.633235 sshd-session[6023]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:12.661420 systemd-logind[1554]: New session 19 of user core. Mar 2 14:30:12.670892 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 2 14:30:13.285551 sshd[6026]: Connection closed by 10.0.0.1 port 39942 Mar 2 14:30:13.292660 sshd-session[6023]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:13.316584 systemd-logind[1554]: Session 19 logged out. Waiting for processes to exit. Mar 2 14:30:13.317878 systemd[1]: sshd@18-10.0.0.9:22-10.0.0.1:39942.service: Deactivated successfully. Mar 2 14:30:13.324877 systemd[1]: session-19.scope: Deactivated successfully. Mar 2 14:30:13.336651 systemd-logind[1554]: Removed session 19. Mar 2 14:30:18.099087 kubelet[2823]: E0302 14:30:18.098516 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:30:18.378735 systemd[1]: Started sshd@19-10.0.0.9:22-10.0.0.1:39956.service - OpenSSH per-connection server daemon (10.0.0.1:39956). Mar 2 14:30:18.724506 sshd[6042]: Accepted publickey for core from 10.0.0.1 port 39956 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:18.730001 sshd-session[6042]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:18.769676 systemd-logind[1554]: New session 20 of user core. Mar 2 14:30:18.797220 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 2 14:30:19.366981 sshd[6045]: Connection closed by 10.0.0.1 port 39956 Mar 2 14:30:19.367836 sshd-session[6042]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:19.376666 systemd[1]: sshd@19-10.0.0.9:22-10.0.0.1:39956.service: Deactivated successfully. Mar 2 14:30:19.396374 systemd[1]: session-20.scope: Deactivated successfully. Mar 2 14:30:19.405097 systemd-logind[1554]: Session 20 logged out. Waiting for processes to exit. Mar 2 14:30:19.411853 systemd-logind[1554]: Removed session 20. Mar 2 14:30:24.455876 systemd[1]: Started sshd@20-10.0.0.9:22-10.0.0.1:37534.service - OpenSSH per-connection server daemon (10.0.0.1:37534). Mar 2 14:30:24.619082 sshd[6092]: Accepted publickey for core from 10.0.0.1 port 37534 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:24.631601 sshd-session[6092]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:24.651580 systemd-logind[1554]: New session 21 of user core. Mar 2 14:30:24.677276 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 2 14:30:25.161754 sshd[6098]: Connection closed by 10.0.0.1 port 37534 Mar 2 14:30:25.163735 sshd-session[6092]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:25.198013 systemd[1]: sshd@20-10.0.0.9:22-10.0.0.1:37534.service: Deactivated successfully. Mar 2 14:30:25.220672 systemd[1]: session-21.scope: Deactivated successfully. Mar 2 14:30:25.243087 systemd-logind[1554]: Session 21 logged out. Waiting for processes to exit. Mar 2 14:30:25.250839 systemd-logind[1554]: Removed session 21. Mar 2 14:30:30.216856 systemd[1]: Started sshd@21-10.0.0.9:22-10.0.0.1:53324.service - OpenSSH per-connection server daemon (10.0.0.1:53324). Mar 2 14:30:30.532073 sshd[6142]: Accepted publickey for core from 10.0.0.1 port 53324 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:30.543145 sshd-session[6142]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:30.608793 systemd-logind[1554]: New session 22 of user core. Mar 2 14:30:30.632179 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 2 14:30:31.210259 sshd[6145]: Connection closed by 10.0.0.1 port 53324 Mar 2 14:30:31.217770 sshd-session[6142]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:31.253838 systemd[1]: sshd@21-10.0.0.9:22-10.0.0.1:53324.service: Deactivated successfully. Mar 2 14:30:31.264659 systemd[1]: session-22.scope: Deactivated successfully. Mar 2 14:30:31.275322 systemd-logind[1554]: Session 22 logged out. Waiting for processes to exit. Mar 2 14:30:31.287265 systemd-logind[1554]: Removed session 22. Mar 2 14:30:36.238205 systemd[1]: Started sshd@22-10.0.0.9:22-10.0.0.1:53340.service - OpenSSH per-connection server daemon (10.0.0.1:53340). Mar 2 14:30:36.471508 sshd[6195]: Accepted publickey for core from 10.0.0.1 port 53340 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:36.479280 sshd-session[6195]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:36.508328 systemd-logind[1554]: New session 23 of user core. Mar 2 14:30:36.533913 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 2 14:30:37.105131 sshd[6198]: Connection closed by 10.0.0.1 port 53340 Mar 2 14:30:37.102651 sshd-session[6195]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:37.143279 systemd[1]: sshd@22-10.0.0.9:22-10.0.0.1:53340.service: Deactivated successfully. Mar 2 14:30:37.163820 systemd[1]: session-23.scope: Deactivated successfully. Mar 2 14:30:37.177975 systemd-logind[1554]: Session 23 logged out. Waiting for processes to exit. Mar 2 14:30:37.182246 systemd-logind[1554]: Removed session 23. Mar 2 14:30:42.120801 kubelet[2823]: E0302 14:30:42.119022 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:30:42.120801 kubelet[2823]: E0302 14:30:42.120394 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:30:42.156664 systemd[1]: Started sshd@23-10.0.0.9:22-10.0.0.1:33384.service - OpenSSH per-connection server daemon (10.0.0.1:33384). Mar 2 14:30:42.417193 sshd[6241]: Accepted publickey for core from 10.0.0.1 port 33384 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:42.428407 sshd-session[6241]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:42.451277 systemd-logind[1554]: New session 24 of user core. Mar 2 14:30:42.470269 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 2 14:30:43.059869 sshd[6244]: Connection closed by 10.0.0.1 port 33384 Mar 2 14:30:43.061393 sshd-session[6241]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:43.075133 systemd[1]: sshd@23-10.0.0.9:22-10.0.0.1:33384.service: Deactivated successfully. Mar 2 14:30:43.083729 systemd[1]: session-24.scope: Deactivated successfully. Mar 2 14:30:43.107078 systemd-logind[1554]: Session 24 logged out. Waiting for processes to exit. Mar 2 14:30:43.112029 systemd-logind[1554]: Removed session 24. Mar 2 14:30:48.141534 systemd[1]: Started sshd@24-10.0.0.9:22-10.0.0.1:33400.service - OpenSSH per-connection server daemon (10.0.0.1:33400). Mar 2 14:30:48.513851 sshd[6258]: Accepted publickey for core from 10.0.0.1 port 33400 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:48.532846 sshd-session[6258]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:48.562712 systemd-logind[1554]: New session 25 of user core. Mar 2 14:30:48.579759 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 2 14:30:49.223397 sshd[6261]: Connection closed by 10.0.0.1 port 33400 Mar 2 14:30:49.221098 sshd-session[6258]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:49.243399 systemd[1]: sshd@24-10.0.0.9:22-10.0.0.1:33400.service: Deactivated successfully. Mar 2 14:30:49.255142 systemd[1]: session-25.scope: Deactivated successfully. Mar 2 14:30:49.264512 systemd-logind[1554]: Session 25 logged out. Waiting for processes to exit. Mar 2 14:30:49.272874 systemd-logind[1554]: Removed session 25. Mar 2 14:30:54.298774 systemd[1]: Started sshd@25-10.0.0.9:22-10.0.0.1:49936.service - OpenSSH per-connection server daemon (10.0.0.1:49936). Mar 2 14:30:54.693362 sshd[6346]: Accepted publickey for core from 10.0.0.1 port 49936 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:30:54.688718 sshd-session[6346]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:30:54.782510 systemd-logind[1554]: New session 26 of user core. Mar 2 14:30:54.839911 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 2 14:30:55.682210 sshd[6349]: Connection closed by 10.0.0.1 port 49936 Mar 2 14:30:55.691696 sshd-session[6346]: pam_unix(sshd:session): session closed for user core Mar 2 14:30:55.714870 systemd[1]: sshd@25-10.0.0.9:22-10.0.0.1:49936.service: Deactivated successfully. Mar 2 14:30:55.727760 systemd[1]: session-26.scope: Deactivated successfully. Mar 2 14:30:55.745513 systemd-logind[1554]: Session 26 logged out. Waiting for processes to exit. Mar 2 14:30:55.761600 systemd-logind[1554]: Removed session 26. Mar 2 14:31:00.101380 kubelet[2823]: E0302 14:31:00.100669 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:31:00.755081 systemd[1]: Started sshd@26-10.0.0.9:22-10.0.0.1:35050.service - OpenSSH per-connection server daemon (10.0.0.1:35050). Mar 2 14:31:01.354008 sshd[6390]: Accepted publickey for core from 10.0.0.1 port 35050 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:01.370832 sshd-session[6390]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:01.399175 systemd-logind[1554]: New session 27 of user core. Mar 2 14:31:01.443118 systemd[1]: Started session-27.scope - Session 27 of User core. Mar 2 14:31:02.404497 sshd[6393]: Connection closed by 10.0.0.1 port 35050 Mar 2 14:31:02.405358 sshd-session[6390]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:02.428065 systemd[1]: sshd@26-10.0.0.9:22-10.0.0.1:35050.service: Deactivated successfully. Mar 2 14:31:02.447919 systemd[1]: session-27.scope: Deactivated successfully. Mar 2 14:31:02.467191 systemd-logind[1554]: Session 27 logged out. Waiting for processes to exit. Mar 2 14:31:02.475660 systemd-logind[1554]: Removed session 27. Mar 2 14:31:07.474839 systemd[1]: Started sshd@27-10.0.0.9:22-10.0.0.1:35060.service - OpenSSH per-connection server daemon (10.0.0.1:35060). Mar 2 14:31:07.722201 sshd[6431]: Accepted publickey for core from 10.0.0.1 port 35060 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:07.734350 sshd-session[6431]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:07.783582 systemd-logind[1554]: New session 28 of user core. Mar 2 14:31:07.809890 systemd[1]: Started session-28.scope - Session 28 of User core. Mar 2 14:31:08.115234 kubelet[2823]: E0302 14:31:08.111038 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:31:08.632198 sshd[6437]: Connection closed by 10.0.0.1 port 35060 Mar 2 14:31:08.628185 sshd-session[6431]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:08.674577 systemd[1]: sshd@27-10.0.0.9:22-10.0.0.1:35060.service: Deactivated successfully. Mar 2 14:31:08.695422 systemd[1]: session-28.scope: Deactivated successfully. Mar 2 14:31:08.722482 systemd-logind[1554]: Session 28 logged out. Waiting for processes to exit. Mar 2 14:31:08.731844 systemd-logind[1554]: Removed session 28. Mar 2 14:31:13.113067 kubelet[2823]: E0302 14:31:13.105856 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:31:13.731992 systemd[1]: Started sshd@28-10.0.0.9:22-10.0.0.1:50540.service - OpenSSH per-connection server daemon (10.0.0.1:50540). Mar 2 14:31:13.953963 sshd[6451]: Accepted publickey for core from 10.0.0.1 port 50540 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:13.957844 sshd-session[6451]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:14.011866 systemd-logind[1554]: New session 29 of user core. Mar 2 14:31:14.047796 systemd[1]: Started session-29.scope - Session 29 of User core. Mar 2 14:31:15.140783 sshd[6454]: Connection closed by 10.0.0.1 port 50540 Mar 2 14:31:15.141375 sshd-session[6451]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:15.149817 systemd[1]: sshd@28-10.0.0.9:22-10.0.0.1:50540.service: Deactivated successfully. Mar 2 14:31:15.170212 systemd[1]: session-29.scope: Deactivated successfully. Mar 2 14:31:15.180410 systemd-logind[1554]: Session 29 logged out. Waiting for processes to exit. Mar 2 14:31:15.189551 systemd-logind[1554]: Removed session 29. Mar 2 14:31:20.184680 systemd[1]: Started sshd@29-10.0.0.9:22-10.0.0.1:45256.service - OpenSSH per-connection server daemon (10.0.0.1:45256). Mar 2 14:31:20.462171 sshd[6491]: Accepted publickey for core from 10.0.0.1 port 45256 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:20.468904 sshd-session[6491]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:20.504203 systemd-logind[1554]: New session 30 of user core. Mar 2 14:31:20.518833 systemd[1]: Started session-30.scope - Session 30 of User core. Mar 2 14:31:21.606102 sshd[6494]: Connection closed by 10.0.0.1 port 45256 Mar 2 14:31:21.603087 sshd-session[6491]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:21.639693 systemd[1]: sshd@29-10.0.0.9:22-10.0.0.1:45256.service: Deactivated successfully. Mar 2 14:31:21.648580 systemd[1]: session-30.scope: Deactivated successfully. Mar 2 14:31:21.660362 systemd-logind[1554]: Session 30 logged out. Waiting for processes to exit. Mar 2 14:31:21.691875 systemd[1]: Started sshd@30-10.0.0.9:22-10.0.0.1:45268.service - OpenSSH per-connection server daemon (10.0.0.1:45268). Mar 2 14:31:21.714747 systemd-logind[1554]: Removed session 30. Mar 2 14:31:22.032908 sshd[6509]: Accepted publickey for core from 10.0.0.1 port 45268 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:22.042377 sshd-session[6509]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:22.070635 systemd-logind[1554]: New session 31 of user core. Mar 2 14:31:22.097118 systemd[1]: Started session-31.scope - Session 31 of User core. Mar 2 14:31:23.355224 sshd[6512]: Connection closed by 10.0.0.1 port 45268 Mar 2 14:31:23.349346 sshd-session[6509]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:23.437143 systemd[1]: sshd@30-10.0.0.9:22-10.0.0.1:45268.service: Deactivated successfully. Mar 2 14:31:23.463653 systemd[1]: session-31.scope: Deactivated successfully. Mar 2 14:31:23.472675 systemd-logind[1554]: Session 31 logged out. Waiting for processes to exit. Mar 2 14:31:23.484238 systemd[1]: Started sshd@31-10.0.0.9:22-10.0.0.1:45274.service - OpenSSH per-connection server daemon (10.0.0.1:45274). Mar 2 14:31:23.499501 systemd-logind[1554]: Removed session 31. Mar 2 14:31:23.801592 sshd[6548]: Accepted publickey for core from 10.0.0.1 port 45274 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:23.807547 sshd-session[6548]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:23.843383 systemd-logind[1554]: New session 32 of user core. Mar 2 14:31:23.868582 systemd[1]: Started session-32.scope - Session 32 of User core. Mar 2 14:31:24.768812 sshd[6551]: Connection closed by 10.0.0.1 port 45274 Mar 2 14:31:24.772407 sshd-session[6548]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:24.857901 systemd[1]: sshd@31-10.0.0.9:22-10.0.0.1:45274.service: Deactivated successfully. Mar 2 14:31:24.870490 systemd[1]: session-32.scope: Deactivated successfully. Mar 2 14:31:24.879571 systemd-logind[1554]: Session 32 logged out. Waiting for processes to exit. Mar 2 14:31:24.915781 systemd-logind[1554]: Removed session 32. Mar 2 14:31:28.107411 kubelet[2823]: E0302 14:31:28.105280 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:31:29.827039 systemd[1]: Started sshd@32-10.0.0.9:22-10.0.0.1:45282.service - OpenSSH per-connection server daemon (10.0.0.1:45282). Mar 2 14:31:30.114151 sshd[6588]: Accepted publickey for core from 10.0.0.1 port 45282 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:30.123103 sshd-session[6588]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:30.166810 systemd-logind[1554]: New session 33 of user core. Mar 2 14:31:30.196679 systemd[1]: Started session-33.scope - Session 33 of User core. Mar 2 14:31:30.736872 sshd[6591]: Connection closed by 10.0.0.1 port 45282 Mar 2 14:31:30.737993 sshd-session[6588]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:30.751029 systemd[1]: sshd@32-10.0.0.9:22-10.0.0.1:45282.service: Deactivated successfully. Mar 2 14:31:30.757397 systemd[1]: session-33.scope: Deactivated successfully. Mar 2 14:31:30.777009 systemd-logind[1554]: Session 33 logged out. Waiting for processes to exit. Mar 2 14:31:30.779934 systemd-logind[1554]: Removed session 33. Mar 2 14:31:35.846546 systemd[1]: Started sshd@33-10.0.0.9:22-10.0.0.1:49232.service - OpenSSH per-connection server daemon (10.0.0.1:49232). Mar 2 14:31:36.055675 sshd[6629]: Accepted publickey for core from 10.0.0.1 port 49232 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:36.063618 sshd-session[6629]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:36.115389 systemd-logind[1554]: New session 34 of user core. Mar 2 14:31:36.128087 systemd[1]: Started session-34.scope - Session 34 of User core. Mar 2 14:31:36.618140 sshd[6632]: Connection closed by 10.0.0.1 port 49232 Mar 2 14:31:36.621621 sshd-session[6629]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:36.633738 systemd[1]: sshd@33-10.0.0.9:22-10.0.0.1:49232.service: Deactivated successfully. Mar 2 14:31:36.642252 systemd[1]: session-34.scope: Deactivated successfully. Mar 2 14:31:36.653019 systemd-logind[1554]: Session 34 logged out. Waiting for processes to exit. Mar 2 14:31:36.660166 systemd-logind[1554]: Removed session 34. Mar 2 14:31:41.683265 systemd[1]: Started sshd@34-10.0.0.9:22-10.0.0.1:32900.service - OpenSSH per-connection server daemon (10.0.0.1:32900). Mar 2 14:31:41.940364 sshd[6645]: Accepted publickey for core from 10.0.0.1 port 32900 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:41.942390 sshd-session[6645]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:41.966683 systemd-logind[1554]: New session 35 of user core. Mar 2 14:31:41.997349 systemd[1]: Started session-35.scope - Session 35 of User core. Mar 2 14:31:42.519369 sshd[6648]: Connection closed by 10.0.0.1 port 32900 Mar 2 14:31:42.521744 sshd-session[6645]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:42.535999 systemd[1]: sshd@34-10.0.0.9:22-10.0.0.1:32900.service: Deactivated successfully. Mar 2 14:31:42.547627 systemd[1]: session-35.scope: Deactivated successfully. Mar 2 14:31:42.566722 systemd-logind[1554]: Session 35 logged out. Waiting for processes to exit. Mar 2 14:31:42.568796 systemd-logind[1554]: Removed session 35. Mar 2 14:31:45.102034 kubelet[2823]: E0302 14:31:45.099725 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:31:47.585360 systemd[1]: Started sshd@35-10.0.0.9:22-10.0.0.1:32906.service - OpenSSH per-connection server daemon (10.0.0.1:32906). Mar 2 14:31:47.917837 sshd[6676]: Accepted publickey for core from 10.0.0.1 port 32906 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:47.930817 sshd-session[6676]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:47.956513 systemd-logind[1554]: New session 36 of user core. Mar 2 14:31:47.972554 systemd[1]: Started session-36.scope - Session 36 of User core. Mar 2 14:31:48.477517 sshd[6679]: Connection closed by 10.0.0.1 port 32906 Mar 2 14:31:48.480016 sshd-session[6676]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:48.520135 systemd[1]: Started sshd@36-10.0.0.9:22-10.0.0.1:32914.service - OpenSSH per-connection server daemon (10.0.0.1:32914). Mar 2 14:31:48.526343 systemd[1]: sshd@35-10.0.0.9:22-10.0.0.1:32906.service: Deactivated successfully. Mar 2 14:31:48.536762 systemd[1]: session-36.scope: Deactivated successfully. Mar 2 14:31:48.557408 systemd-logind[1554]: Session 36 logged out. Waiting for processes to exit. Mar 2 14:31:48.580265 systemd-logind[1554]: Removed session 36. Mar 2 14:31:49.323225 sshd[6690]: Accepted publickey for core from 10.0.0.1 port 32914 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:49.328840 sshd-session[6690]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:49.364843 systemd-logind[1554]: New session 37 of user core. Mar 2 14:31:49.417129 systemd[1]: Started session-37.scope - Session 37 of User core. Mar 2 14:31:50.715938 sshd[6697]: Connection closed by 10.0.0.1 port 32914 Mar 2 14:31:50.719940 sshd-session[6690]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:50.758240 systemd[1]: Started sshd@37-10.0.0.9:22-10.0.0.1:57600.service - OpenSSH per-connection server daemon (10.0.0.1:57600). Mar 2 14:31:50.765218 systemd[1]: sshd@36-10.0.0.9:22-10.0.0.1:32914.service: Deactivated successfully. Mar 2 14:31:50.776622 systemd[1]: session-37.scope: Deactivated successfully. Mar 2 14:31:50.783897 systemd-logind[1554]: Session 37 logged out. Waiting for processes to exit. Mar 2 14:31:50.799825 systemd-logind[1554]: Removed session 37. Mar 2 14:31:50.965097 sshd[6706]: Accepted publickey for core from 10.0.0.1 port 57600 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:50.968134 sshd-session[6706]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:50.994672 systemd-logind[1554]: New session 38 of user core. Mar 2 14:31:51.028342 systemd[1]: Started session-38.scope - Session 38 of User core. Mar 2 14:31:53.747575 sshd[6712]: Connection closed by 10.0.0.1 port 57600 Mar 2 14:31:53.755231 sshd-session[6706]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:53.783487 systemd[1]: sshd@37-10.0.0.9:22-10.0.0.1:57600.service: Deactivated successfully. Mar 2 14:31:53.795286 systemd[1]: session-38.scope: Deactivated successfully. Mar 2 14:31:53.813476 systemd-logind[1554]: Session 38 logged out. Waiting for processes to exit. Mar 2 14:31:53.830247 systemd[1]: Started sshd@38-10.0.0.9:22-10.0.0.1:57614.service - OpenSSH per-connection server daemon (10.0.0.1:57614). Mar 2 14:31:53.841926 systemd-logind[1554]: Removed session 38. Mar 2 14:31:54.067143 sshd[6811]: Accepted publickey for core from 10.0.0.1 port 57614 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:54.070313 sshd-session[6811]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:54.102623 systemd-logind[1554]: New session 39 of user core. Mar 2 14:31:54.106145 systemd[1]: Started session-39.scope - Session 39 of User core. Mar 2 14:31:56.097647 sshd[6816]: Connection closed by 10.0.0.1 port 57614 Mar 2 14:31:56.102157 sshd-session[6811]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:56.206630 systemd[1]: sshd@38-10.0.0.9:22-10.0.0.1:57614.service: Deactivated successfully. Mar 2 14:31:56.218547 systemd[1]: session-39.scope: Deactivated successfully. Mar 2 14:31:56.221727 systemd-logind[1554]: Session 39 logged out. Waiting for processes to exit. Mar 2 14:31:56.238943 systemd[1]: Started sshd@39-10.0.0.9:22-10.0.0.1:57626.service - OpenSSH per-connection server daemon (10.0.0.1:57626). Mar 2 14:31:56.241910 systemd-logind[1554]: Removed session 39. Mar 2 14:31:56.468217 sshd[6846]: Accepted publickey for core from 10.0.0.1 port 57626 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:31:56.473431 sshd-session[6846]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:31:56.511758 systemd-logind[1554]: New session 40 of user core. Mar 2 14:31:56.525678 systemd[1]: Started session-40.scope - Session 40 of User core. Mar 2 14:31:57.033749 sshd[6856]: Connection closed by 10.0.0.1 port 57626 Mar 2 14:31:57.034606 sshd-session[6846]: pam_unix(sshd:session): session closed for user core Mar 2 14:31:57.064708 systemd[1]: sshd@39-10.0.0.9:22-10.0.0.1:57626.service: Deactivated successfully. Mar 2 14:31:57.069312 systemd[1]: session-40.scope: Deactivated successfully. Mar 2 14:31:57.095754 systemd-logind[1554]: Session 40 logged out. Waiting for processes to exit. Mar 2 14:31:57.130277 systemd-logind[1554]: Removed session 40. Mar 2 14:32:02.043116 containerd[1574]: time="2026-03-02T14:32:02.012675340Z" level=warning msg="container event discarded" container=b45025ecf2a4016256f3b5752dd43b1dc00848de1bbadbe9f725232e23f21851 type=CONTAINER_CREATED_EVENT Mar 2 14:32:02.043116 containerd[1574]: time="2026-03-02T14:32:02.042969500Z" level=warning msg="container event discarded" container=b45025ecf2a4016256f3b5752dd43b1dc00848de1bbadbe9f725232e23f21851 type=CONTAINER_STARTED_EVENT Mar 2 14:32:02.121651 containerd[1574]: time="2026-03-02T14:32:02.107816597Z" level=warning msg="container event discarded" container=962f50155d5c0cac4a96191acea6e5904eb1a472dffa79372b30ceb105a62329 type=CONTAINER_CREATED_EVENT Mar 2 14:32:02.121651 containerd[1574]: time="2026-03-02T14:32:02.107886888Z" level=warning msg="container event discarded" container=962f50155d5c0cac4a96191acea6e5904eb1a472dffa79372b30ceb105a62329 type=CONTAINER_STARTED_EVENT Mar 2 14:32:02.121651 containerd[1574]: time="2026-03-02T14:32:02.107901956Z" level=warning msg="container event discarded" container=7e1777010bed87983c41956462aa5590efc6a2e54f2b0cbb9c18c3c8381fcc41 type=CONTAINER_CREATED_EVENT Mar 2 14:32:02.121651 containerd[1574]: time="2026-03-02T14:32:02.107913398Z" level=warning msg="container event discarded" container=7e1777010bed87983c41956462aa5590efc6a2e54f2b0cbb9c18c3c8381fcc41 type=CONTAINER_STARTED_EVENT Mar 2 14:32:02.121651 containerd[1574]: time="2026-03-02T14:32:02.107924548Z" level=warning msg="container event discarded" container=b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f type=CONTAINER_CREATED_EVENT Mar 2 14:32:02.122402 containerd[1574]: time="2026-03-02T14:32:02.122307608Z" level=warning msg="container event discarded" container=833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947 type=CONTAINER_CREATED_EVENT Mar 2 14:32:02.129863 systemd[1]: Started sshd@40-10.0.0.9:22-10.0.0.1:41800.service - OpenSSH per-connection server daemon (10.0.0.1:41800). Mar 2 14:32:02.197896 containerd[1574]: time="2026-03-02T14:32:02.197642661Z" level=warning msg="container event discarded" container=3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc type=CONTAINER_CREATED_EVENT Mar 2 14:32:02.323316 sshd[6869]: Accepted publickey for core from 10.0.0.1 port 41800 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:32:02.332329 sshd-session[6869]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:32:02.353187 systemd-logind[1554]: New session 41 of user core. Mar 2 14:32:02.371550 systemd[1]: Started session-41.scope - Session 41 of User core. Mar 2 14:32:02.416399 containerd[1574]: time="2026-03-02T14:32:02.415407209Z" level=warning msg="container event discarded" container=b7f7cde178bff1e5c7a1395a3046e7ab9193000243fa3530b3730e54ad9a462f type=CONTAINER_STARTED_EVENT Mar 2 14:32:02.496804 containerd[1574]: time="2026-03-02T14:32:02.496713373Z" level=warning msg="container event discarded" container=833ad0c8afbc8d73ef364e4a23ad30c82dcc0c27c7059317dc3fe4d3aa027947 type=CONTAINER_STARTED_EVENT Mar 2 14:32:02.572403 containerd[1574]: time="2026-03-02T14:32:02.566287283Z" level=warning msg="container event discarded" container=3ae18a213ee12c61f414daf1d8e20ce04c0846f5d352f1bf2eec4b18b92d03dc type=CONTAINER_STARTED_EVENT Mar 2 14:32:02.916522 sshd[6872]: Connection closed by 10.0.0.1 port 41800 Mar 2 14:32:02.917753 sshd-session[6869]: pam_unix(sshd:session): session closed for user core Mar 2 14:32:02.931151 systemd[1]: sshd@40-10.0.0.9:22-10.0.0.1:41800.service: Deactivated successfully. Mar 2 14:32:02.933965 systemd[1]: session-41.scope: Deactivated successfully. Mar 2 14:32:02.948201 systemd-logind[1554]: Session 41 logged out. Waiting for processes to exit. Mar 2 14:32:02.958543 systemd-logind[1554]: Removed session 41. Mar 2 14:32:07.949274 systemd[1]: Started sshd@41-10.0.0.9:22-10.0.0.1:41812.service - OpenSSH per-connection server daemon (10.0.0.1:41812). Mar 2 14:32:08.369135 sshd[6907]: Accepted publickey for core from 10.0.0.1 port 41812 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:32:08.375749 sshd-session[6907]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:32:08.397590 systemd-logind[1554]: New session 42 of user core. Mar 2 14:32:08.402830 systemd[1]: Started session-42.scope - Session 42 of User core. Mar 2 14:32:08.952206 sshd[6910]: Connection closed by 10.0.0.1 port 41812 Mar 2 14:32:08.953720 sshd-session[6907]: pam_unix(sshd:session): session closed for user core Mar 2 14:32:08.996971 systemd[1]: sshd@41-10.0.0.9:22-10.0.0.1:41812.service: Deactivated successfully. Mar 2 14:32:09.018532 systemd[1]: session-42.scope: Deactivated successfully. Mar 2 14:32:09.042529 systemd-logind[1554]: Session 42 logged out. Waiting for processes to exit. Mar 2 14:32:09.051922 systemd-logind[1554]: Removed session 42. Mar 2 14:32:10.109170 kubelet[2823]: E0302 14:32:10.106928 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:32:11.129517 kubelet[2823]: E0302 14:32:11.100910 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:32:13.105499 kubelet[2823]: E0302 14:32:13.102979 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:32:14.016504 systemd[1]: Started sshd@42-10.0.0.9:22-10.0.0.1:43858.service - OpenSSH per-connection server daemon (10.0.0.1:43858). Mar 2 14:32:14.196930 sshd[6944]: Accepted publickey for core from 10.0.0.1 port 43858 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:32:14.205819 sshd-session[6944]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:32:14.246026 systemd-logind[1554]: New session 43 of user core. Mar 2 14:32:14.271349 systemd[1]: Started session-43.scope - Session 43 of User core. Mar 2 14:32:14.813546 sshd[6947]: Connection closed by 10.0.0.1 port 43858 Mar 2 14:32:14.811540 sshd-session[6944]: pam_unix(sshd:session): session closed for user core Mar 2 14:32:14.830089 systemd[1]: sshd@42-10.0.0.9:22-10.0.0.1:43858.service: Deactivated successfully. Mar 2 14:32:14.845715 systemd[1]: session-43.scope: Deactivated successfully. Mar 2 14:32:14.851780 systemd-logind[1554]: Session 43 logged out. Waiting for processes to exit. Mar 2 14:32:14.858852 systemd-logind[1554]: Removed session 43. Mar 2 14:32:19.859522 systemd[1]: Started sshd@43-10.0.0.9:22-10.0.0.1:43870.service - OpenSSH per-connection server daemon (10.0.0.1:43870). Mar 2 14:32:20.035277 sshd[6977]: Accepted publickey for core from 10.0.0.1 port 43870 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:32:20.045843 sshd-session[6977]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:32:20.106858 systemd-logind[1554]: New session 44 of user core. Mar 2 14:32:20.118195 systemd[1]: Started session-44.scope - Session 44 of User core. Mar 2 14:32:20.477772 sshd[6980]: Connection closed by 10.0.0.1 port 43870 Mar 2 14:32:20.476869 sshd-session[6977]: pam_unix(sshd:session): session closed for user core Mar 2 14:32:20.499556 systemd-logind[1554]: Session 44 logged out. Waiting for processes to exit. Mar 2 14:32:20.500588 systemd[1]: sshd@43-10.0.0.9:22-10.0.0.1:43870.service: Deactivated successfully. Mar 2 14:32:20.524822 systemd[1]: session-44.scope: Deactivated successfully. Mar 2 14:32:20.537875 systemd-logind[1554]: Removed session 44. Mar 2 14:32:21.506103 containerd[1574]: time="2026-03-02T14:32:21.505954203Z" level=warning msg="container event discarded" container=f59e24a7603223a062474f238d0a467dc0cea35402d78bbd96eb1649aa3461e2 type=CONTAINER_CREATED_EVENT Mar 2 14:32:21.506103 containerd[1574]: time="2026-03-02T14:32:21.506069869Z" level=warning msg="container event discarded" container=f59e24a7603223a062474f238d0a467dc0cea35402d78bbd96eb1649aa3461e2 type=CONTAINER_STARTED_EVENT Mar 2 14:32:21.595406 containerd[1574]: time="2026-03-02T14:32:21.595328538Z" level=warning msg="container event discarded" container=8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43 type=CONTAINER_CREATED_EVENT Mar 2 14:32:21.979525 containerd[1574]: time="2026-03-02T14:32:21.977510056Z" level=warning msg="container event discarded" container=8c739fe4b4e0f0d189ece8b7d438e463f96b691cb505f456a911f8b401920f43 type=CONTAINER_STARTED_EVENT Mar 2 14:32:22.040151 containerd[1574]: time="2026-03-02T14:32:22.040033732Z" level=warning msg="container event discarded" container=d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7 type=CONTAINER_CREATED_EVENT Mar 2 14:32:22.040151 containerd[1574]: time="2026-03-02T14:32:22.040106808Z" level=warning msg="container event discarded" container=d41db90148cd5085b295b6d787f0699d31df584f2111c0079b544879f3bd91a7 type=CONTAINER_STARTED_EVENT Mar 2 14:32:24.105960 kubelet[2823]: E0302 14:32:24.103584 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:32:25.538527 systemd[1]: Started sshd@44-10.0.0.9:22-10.0.0.1:42892.service - OpenSSH per-connection server daemon (10.0.0.1:42892). Mar 2 14:32:25.869625 sshd[7020]: Accepted publickey for core from 10.0.0.1 port 42892 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:32:25.894061 sshd-session[7020]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:32:25.958537 systemd-logind[1554]: New session 45 of user core. Mar 2 14:32:25.972490 systemd[1]: Started session-45.scope - Session 45 of User core. Mar 2 14:32:27.208167 sshd[7023]: Connection closed by 10.0.0.1 port 42892 Mar 2 14:32:27.210237 sshd-session[7020]: pam_unix(sshd:session): session closed for user core Mar 2 14:32:27.242277 systemd-logind[1554]: Session 45 logged out. Waiting for processes to exit. Mar 2 14:32:27.246430 systemd[1]: sshd@44-10.0.0.9:22-10.0.0.1:42892.service: Deactivated successfully. Mar 2 14:32:27.257754 systemd[1]: session-45.scope: Deactivated successfully. Mar 2 14:32:27.272683 systemd-logind[1554]: Removed session 45. Mar 2 14:32:27.834990 containerd[1574]: time="2026-03-02T14:32:27.825782204Z" level=warning msg="container event discarded" container=0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a type=CONTAINER_CREATED_EVENT Mar 2 14:32:27.954664 containerd[1574]: time="2026-03-02T14:32:27.954381276Z" level=warning msg="container event discarded" container=0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a type=CONTAINER_STARTED_EVENT Mar 2 14:32:32.267686 systemd[1]: Started sshd@45-10.0.0.9:22-10.0.0.1:60648.service - OpenSSH per-connection server daemon (10.0.0.1:60648). Mar 2 14:32:32.507489 sshd[7063]: Accepted publickey for core from 10.0.0.1 port 60648 ssh2: RSA SHA256:YvdBDTdEI1lli8iGgRc26R2mJamvNBJNeePgmjt42C0 Mar 2 14:32:32.508791 sshd-session[7063]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 14:32:32.529552 systemd-logind[1554]: New session 46 of user core. Mar 2 14:32:32.546418 systemd[1]: Started session-46.scope - Session 46 of User core. Mar 2 14:32:32.971669 containerd[1574]: time="2026-03-02T14:32:32.968310736Z" level=warning msg="container event discarded" container=0871b1361ef8b289e94b2be2048f5ab3d2c30b8e22b17c3164ad8d2d6cff2e3a type=CONTAINER_STOPPED_EVENT Mar 2 14:32:33.010770 sshd[7066]: Connection closed by 10.0.0.1 port 60648 Mar 2 14:32:33.010272 sshd-session[7063]: pam_unix(sshd:session): session closed for user core Mar 2 14:32:33.027554 systemd[1]: sshd@45-10.0.0.9:22-10.0.0.1:60648.service: Deactivated successfully. Mar 2 14:32:33.038766 systemd[1]: session-46.scope: Deactivated successfully. Mar 2 14:32:33.047054 systemd-logind[1554]: Session 46 logged out. Waiting for processes to exit. Mar 2 14:32:33.056146 systemd-logind[1554]: Removed session 46. Mar 2 14:32:33.102507 kubelet[2823]: E0302 14:32:33.100912 2823 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 14:32:33.466376 containerd[1574]: time="2026-03-02T14:32:33.455629640Z" level=warning msg="container event discarded" container=0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812 type=CONTAINER_CREATED_EVENT Mar 2 14:32:33.676160 containerd[1574]: time="2026-03-02T14:32:33.675223828Z" level=warning msg="container event discarded" container=0e122307c4f89994294661f3f5ece6fae738c6f1c416ed881cdaf03839348812 type=CONTAINER_STARTED_EVENT