Apr 14 13:31:12.880665 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Mon Apr 13 18:40:27 -00 2026 Apr 14 13:31:12.880684 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c1ba97db2f6278922cfc5bd0ca74b4bb573fca2c3aed19c121a34271e693e156 Apr 14 13:31:12.880694 kernel: BIOS-provided physical RAM map: Apr 14 13:31:12.880700 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Apr 14 13:31:12.880705 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Apr 14 13:31:12.880710 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Apr 14 13:31:12.880738 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Apr 14 13:31:12.880747 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Apr 14 13:31:12.880755 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Apr 14 13:31:12.880763 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Apr 14 13:31:12.880768 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Apr 14 13:31:12.880773 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Apr 14 13:31:12.880778 kernel: NX (Execute Disable) protection: active Apr 14 13:31:12.880783 kernel: APIC: Static calls initialized Apr 14 13:31:12.880790 kernel: SMBIOS 2.8 present. Apr 14 13:31:12.880797 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Apr 14 13:31:12.880803 kernel: Hypervisor detected: KVM Apr 14 13:31:12.880808 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Apr 14 13:31:12.880814 kernel: kvm-clock: using sched offset of 3437183431 cycles Apr 14 13:31:12.880820 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Apr 14 13:31:12.880826 kernel: tsc: Detected 2793.438 MHz processor Apr 14 13:31:12.880832 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Apr 14 13:31:12.880838 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Apr 14 13:31:12.880843 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x10000000000 Apr 14 13:31:12.880851 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Apr 14 13:31:12.880856 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Apr 14 13:31:12.880862 kernel: Using GB pages for direct mapping Apr 14 13:31:12.880868 kernel: ACPI: Early table checksum verification disabled Apr 14 13:31:12.880874 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Apr 14 13:31:12.880879 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 14 13:31:12.880885 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 14 13:31:12.880891 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 14 13:31:12.880895 kernel: ACPI: FACS 0x000000009CFE0000 000040 Apr 14 13:31:12.880901 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 14 13:31:12.880906 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 14 13:31:12.880911 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 14 13:31:12.880915 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 14 13:31:12.880920 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Apr 14 13:31:12.880925 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Apr 14 13:31:12.880929 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Apr 14 13:31:12.880936 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Apr 14 13:31:12.880943 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Apr 14 13:31:12.880948 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Apr 14 13:31:12.880953 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Apr 14 13:31:12.880957 kernel: No NUMA configuration found Apr 14 13:31:12.880962 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Apr 14 13:31:12.880967 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Apr 14 13:31:12.880974 kernel: Zone ranges: Apr 14 13:31:12.880979 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Apr 14 13:31:12.880984 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Apr 14 13:31:12.880988 kernel: Normal empty Apr 14 13:31:12.880993 kernel: Movable zone start for each node Apr 14 13:31:12.880998 kernel: Early memory node ranges Apr 14 13:31:12.881003 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Apr 14 13:31:12.881008 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Apr 14 13:31:12.881013 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Apr 14 13:31:12.881018 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Apr 14 13:31:12.881024 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Apr 14 13:31:12.881029 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Apr 14 13:31:12.881034 kernel: ACPI: PM-Timer IO Port: 0x608 Apr 14 13:31:12.881039 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Apr 14 13:31:12.881044 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Apr 14 13:31:12.881049 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Apr 14 13:31:12.881054 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Apr 14 13:31:12.881058 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Apr 14 13:31:12.881063 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Apr 14 13:31:12.881070 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Apr 14 13:31:12.881075 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Apr 14 13:31:12.881079 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Apr 14 13:31:12.881084 kernel: TSC deadline timer available Apr 14 13:31:12.881089 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Apr 14 13:31:12.881094 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Apr 14 13:31:12.881099 kernel: kvm-guest: KVM setup pv remote TLB flush Apr 14 13:31:12.881104 kernel: kvm-guest: setup PV sched yield Apr 14 13:31:12.881109 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Apr 14 13:31:12.881115 kernel: Booting paravirtualized kernel on KVM Apr 14 13:31:12.881120 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Apr 14 13:31:12.881126 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Apr 14 13:31:12.881135 kernel: percpu: Embedded 57 pages/cpu s196328 r8192 d28952 u524288 Apr 14 13:31:12.881144 kernel: pcpu-alloc: s196328 r8192 d28952 u524288 alloc=1*2097152 Apr 14 13:31:12.881151 kernel: pcpu-alloc: [0] 0 1 2 3 Apr 14 13:31:12.881159 kernel: kvm-guest: PV spinlocks enabled Apr 14 13:31:12.881168 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Apr 14 13:31:12.881178 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c1ba97db2f6278922cfc5bd0ca74b4bb573fca2c3aed19c121a34271e693e156 Apr 14 13:31:12.881190 kernel: random: crng init done Apr 14 13:31:12.881195 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 14 13:31:12.881200 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 14 13:31:12.881204 kernel: Fallback order for Node 0: 0 Apr 14 13:31:12.881209 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Apr 14 13:31:12.881214 kernel: Policy zone: DMA32 Apr 14 13:31:12.881245 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 14 13:31:12.881250 kernel: Memory: 2433652K/2571752K available (12288K kernel code, 2288K rwdata, 22752K rodata, 42896K init, 2300K bss, 137896K reserved, 0K cma-reserved) Apr 14 13:31:12.881258 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Apr 14 13:31:12.881262 kernel: ftrace: allocating 37996 entries in 149 pages Apr 14 13:31:12.881268 kernel: ftrace: allocated 149 pages with 4 groups Apr 14 13:31:12.881273 kernel: Dynamic Preempt: voluntary Apr 14 13:31:12.881278 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 14 13:31:12.881283 kernel: rcu: RCU event tracing is enabled. Apr 14 13:31:12.881289 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Apr 14 13:31:12.881294 kernel: Trampoline variant of Tasks RCU enabled. Apr 14 13:31:12.881299 kernel: Rude variant of Tasks RCU enabled. Apr 14 13:31:12.881304 kernel: Tracing variant of Tasks RCU enabled. Apr 14 13:31:12.881310 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 14 13:31:12.881315 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Apr 14 13:31:12.881320 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Apr 14 13:31:12.881325 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 14 13:31:12.881330 kernel: Console: colour VGA+ 80x25 Apr 14 13:31:12.881334 kernel: printk: console [ttyS0] enabled Apr 14 13:31:12.881339 kernel: ACPI: Core revision 20230628 Apr 14 13:31:12.881345 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Apr 14 13:31:12.881350 kernel: APIC: Switch to symmetric I/O mode setup Apr 14 13:31:12.881356 kernel: x2apic enabled Apr 14 13:31:12.881361 kernel: APIC: Switched APIC routing to: physical x2apic Apr 14 13:31:12.881366 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Apr 14 13:31:12.881371 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Apr 14 13:31:12.881376 kernel: kvm-guest: setup PV IPIs Apr 14 13:31:12.881381 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Apr 14 13:31:12.881386 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x284409db922, max_idle_ns: 440795228871 ns Apr 14 13:31:12.881398 kernel: Calibrating delay loop (skipped) preset value.. 5586.87 BogoMIPS (lpj=2793438) Apr 14 13:31:12.881403 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Apr 14 13:31:12.881409 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Apr 14 13:31:12.881414 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Apr 14 13:31:12.881421 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Apr 14 13:31:12.881426 kernel: Spectre V2 : Mitigation: Retpolines Apr 14 13:31:12.881432 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Apr 14 13:31:12.881437 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Apr 14 13:31:12.881443 kernel: RETBleed: Vulnerable Apr 14 13:31:12.881449 kernel: Speculative Store Bypass: Vulnerable Apr 14 13:31:12.881455 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Apr 14 13:31:12.881460 kernel: GDS: Unknown: Dependent on hypervisor status Apr 14 13:31:12.881466 kernel: active return thunk: its_return_thunk Apr 14 13:31:12.881471 kernel: ITS: Mitigation: Aligned branch/return thunks Apr 14 13:31:12.881476 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Apr 14 13:31:12.881482 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Apr 14 13:31:12.881487 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Apr 14 13:31:12.881493 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Apr 14 13:31:12.881499 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Apr 14 13:31:12.881505 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Apr 14 13:31:12.881510 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Apr 14 13:31:12.881516 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Apr 14 13:31:12.881521 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Apr 14 13:31:12.881526 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Apr 14 13:31:12.881532 kernel: x86/fpu: Enabled xstate features 0xe7, context size is 2432 bytes, using 'compacted' format. Apr 14 13:31:12.881537 kernel: Freeing SMP alternatives memory: 32K Apr 14 13:31:12.881543 kernel: pid_max: default: 32768 minimum: 301 Apr 14 13:31:12.881549 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 14 13:31:12.881555 kernel: landlock: Up and running. Apr 14 13:31:12.881560 kernel: SELinux: Initializing. Apr 14 13:31:12.881566 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 14 13:31:12.881571 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 14 13:31:12.881577 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8370C CPU @ 2.80GHz (family: 0x6, model: 0x6a, stepping: 0x6) Apr 14 13:31:12.881582 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 14 13:31:12.881588 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 14 13:31:12.881594 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 14 13:31:12.881600 kernel: Performance Events: unsupported p6 CPU model 106 no PMU driver, software events only. Apr 14 13:31:12.881605 kernel: signal: max sigframe size: 3632 Apr 14 13:31:12.881610 kernel: rcu: Hierarchical SRCU implementation. Apr 14 13:31:12.881616 kernel: rcu: Max phase no-delay instances is 400. Apr 14 13:31:12.881621 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Apr 14 13:31:12.881627 kernel: smp: Bringing up secondary CPUs ... Apr 14 13:31:12.881632 kernel: smpboot: x86: Booting SMP configuration: Apr 14 13:31:12.881637 kernel: .... node #0, CPUs: #1 #2 #3 Apr 14 13:31:12.881644 kernel: smp: Brought up 1 node, 4 CPUs Apr 14 13:31:12.881649 kernel: smpboot: Max logical packages: 1 Apr 14 13:31:12.881655 kernel: smpboot: Total of 4 processors activated (22347.50 BogoMIPS) Apr 14 13:31:12.881660 kernel: devtmpfs: initialized Apr 14 13:31:12.881666 kernel: x86/mm: Memory block size: 128MB Apr 14 13:31:12.881671 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 14 13:31:12.881677 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Apr 14 13:31:12.881682 kernel: pinctrl core: initialized pinctrl subsystem Apr 14 13:31:12.881687 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 14 13:31:12.881693 kernel: audit: initializing netlink subsys (disabled) Apr 14 13:31:12.881699 kernel: audit: type=2000 audit(1776173471.925:1): state=initialized audit_enabled=0 res=1 Apr 14 13:31:12.881704 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 14 13:31:12.881710 kernel: thermal_sys: Registered thermal governor 'user_space' Apr 14 13:31:12.881731 kernel: cpuidle: using governor menu Apr 14 13:31:12.881741 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 14 13:31:12.881750 kernel: dca service started, version 1.12.1 Apr 14 13:31:12.881755 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Apr 14 13:31:12.881763 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Apr 14 13:31:12.881771 kernel: PCI: Using configuration type 1 for base access Apr 14 13:31:12.881776 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Apr 14 13:31:12.881782 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 14 13:31:12.881787 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Apr 14 13:31:12.881793 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 14 13:31:12.881798 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Apr 14 13:31:12.881803 kernel: ACPI: Added _OSI(Module Device) Apr 14 13:31:12.881809 kernel: ACPI: Added _OSI(Processor Device) Apr 14 13:31:12.881814 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 14 13:31:12.881821 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 14 13:31:12.881827 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Apr 14 13:31:12.881832 kernel: ACPI: Interpreter enabled Apr 14 13:31:12.881837 kernel: ACPI: PM: (supports S0 S3 S5) Apr 14 13:31:12.881843 kernel: ACPI: Using IOAPIC for interrupt routing Apr 14 13:31:12.881848 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Apr 14 13:31:12.881854 kernel: PCI: Using E820 reservations for host bridge windows Apr 14 13:31:12.881859 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Apr 14 13:31:12.881864 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 14 13:31:12.881980 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 14 13:31:12.882043 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Apr 14 13:31:12.882099 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Apr 14 13:31:12.882106 kernel: PCI host bridge to bus 0000:00 Apr 14 13:31:12.882183 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Apr 14 13:31:12.882291 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Apr 14 13:31:12.882344 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Apr 14 13:31:12.882396 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Apr 14 13:31:12.882446 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Apr 14 13:31:12.882495 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Apr 14 13:31:12.882543 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 14 13:31:12.882611 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Apr 14 13:31:12.882678 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Apr 14 13:31:12.882767 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Apr 14 13:31:12.882825 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Apr 14 13:31:12.882881 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Apr 14 13:31:12.882935 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Apr 14 13:31:12.882995 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Apr 14 13:31:12.883052 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Apr 14 13:31:12.883107 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Apr 14 13:31:12.883185 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Apr 14 13:31:12.883293 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Apr 14 13:31:12.883352 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Apr 14 13:31:12.883408 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Apr 14 13:31:12.883463 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Apr 14 13:31:12.883525 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Apr 14 13:31:12.883581 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Apr 14 13:31:12.883639 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Apr 14 13:31:12.883693 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Apr 14 13:31:12.883778 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Apr 14 13:31:12.883839 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Apr 14 13:31:12.883895 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Apr 14 13:31:12.883954 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Apr 14 13:31:12.884013 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Apr 14 13:31:12.884067 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Apr 14 13:31:12.884127 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Apr 14 13:31:12.884207 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Apr 14 13:31:12.884215 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Apr 14 13:31:12.884259 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Apr 14 13:31:12.884265 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Apr 14 13:31:12.884271 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Apr 14 13:31:12.884279 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Apr 14 13:31:12.884284 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Apr 14 13:31:12.884290 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Apr 14 13:31:12.884295 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Apr 14 13:31:12.884300 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Apr 14 13:31:12.884306 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Apr 14 13:31:12.884311 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Apr 14 13:31:12.884317 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Apr 14 13:31:12.884323 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Apr 14 13:31:12.884330 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Apr 14 13:31:12.884336 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Apr 14 13:31:12.884342 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Apr 14 13:31:12.884347 kernel: iommu: Default domain type: Translated Apr 14 13:31:12.884353 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Apr 14 13:31:12.884358 kernel: PCI: Using ACPI for IRQ routing Apr 14 13:31:12.884364 kernel: PCI: pci_cache_line_size set to 64 bytes Apr 14 13:31:12.884369 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Apr 14 13:31:12.884375 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Apr 14 13:31:12.884438 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Apr 14 13:31:12.884493 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Apr 14 13:31:12.884548 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Apr 14 13:31:12.884555 kernel: vgaarb: loaded Apr 14 13:31:12.884561 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Apr 14 13:31:12.884566 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Apr 14 13:31:12.884571 kernel: clocksource: Switched to clocksource kvm-clock Apr 14 13:31:12.884577 kernel: VFS: Disk quotas dquot_6.6.0 Apr 14 13:31:12.884582 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 14 13:31:12.884590 kernel: pnp: PnP ACPI init Apr 14 13:31:12.884656 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Apr 14 13:31:12.884664 kernel: pnp: PnP ACPI: found 6 devices Apr 14 13:31:12.884670 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Apr 14 13:31:12.884676 kernel: NET: Registered PF_INET protocol family Apr 14 13:31:12.884681 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 14 13:31:12.884687 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 14 13:31:12.884693 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 14 13:31:12.884700 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 14 13:31:12.884706 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 14 13:31:12.884711 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 14 13:31:12.884739 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 14 13:31:12.884750 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 14 13:31:12.884758 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 14 13:31:12.884764 kernel: NET: Registered PF_XDP protocol family Apr 14 13:31:12.884821 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Apr 14 13:31:12.884871 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Apr 14 13:31:12.884922 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Apr 14 13:31:12.884971 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Apr 14 13:31:12.885020 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Apr 14 13:31:12.885069 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Apr 14 13:31:12.885076 kernel: PCI: CLS 0 bytes, default 64 Apr 14 13:31:12.885082 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Apr 14 13:31:12.885088 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x284409db922, max_idle_ns: 440795228871 ns Apr 14 13:31:12.885093 kernel: Initialise system trusted keyrings Apr 14 13:31:12.885101 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 14 13:31:12.885106 kernel: Key type asymmetric registered Apr 14 13:31:12.885112 kernel: Asymmetric key parser 'x509' registered Apr 14 13:31:12.885117 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Apr 14 13:31:12.885123 kernel: io scheduler mq-deadline registered Apr 14 13:31:12.885131 kernel: io scheduler kyber registered Apr 14 13:31:12.885140 kernel: io scheduler bfq registered Apr 14 13:31:12.885149 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Apr 14 13:31:12.885158 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Apr 14 13:31:12.885170 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Apr 14 13:31:12.885180 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Apr 14 13:31:12.885190 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 14 13:31:12.885195 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Apr 14 13:31:12.885201 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Apr 14 13:31:12.885207 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Apr 14 13:31:12.885213 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Apr 14 13:31:12.885317 kernel: rtc_cmos 00:04: RTC can wake from S4 Apr 14 13:31:12.885328 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Apr 14 13:31:12.885378 kernel: rtc_cmos 00:04: registered as rtc0 Apr 14 13:31:12.885430 kernel: rtc_cmos 00:04: setting system clock to 2026-04-14T13:31:12 UTC (1776173472) Apr 14 13:31:12.885481 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Apr 14 13:31:12.885488 kernel: intel_pstate: CPU model not supported Apr 14 13:31:12.885493 kernel: NET: Registered PF_INET6 protocol family Apr 14 13:31:12.885499 kernel: Segment Routing with IPv6 Apr 14 13:31:12.885504 kernel: In-situ OAM (IOAM) with IPv6 Apr 14 13:31:12.885509 kernel: NET: Registered PF_PACKET protocol family Apr 14 13:31:12.885517 kernel: Key type dns_resolver registered Apr 14 13:31:12.885523 kernel: IPI shorthand broadcast: enabled Apr 14 13:31:12.885528 kernel: sched_clock: Marking stable (721072411, 272051083)->(1103138374, -110014880) Apr 14 13:31:12.885534 kernel: registered taskstats version 1 Apr 14 13:31:12.885540 kernel: Loading compiled-in X.509 certificates Apr 14 13:31:12.885545 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: 51221ce98a81ccf90ef3d16403b42695603c5d00' Apr 14 13:31:12.885551 kernel: Key type .fscrypt registered Apr 14 13:31:12.885557 kernel: Key type fscrypt-provisioning registered Apr 14 13:31:12.885562 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 14 13:31:12.885569 kernel: ima: Allocated hash algorithm: sha1 Apr 14 13:31:12.885575 kernel: ima: No architecture policies found Apr 14 13:31:12.885581 kernel: clk: Disabling unused clocks Apr 14 13:31:12.885587 kernel: Freeing unused kernel image (initmem) memory: 42896K Apr 14 13:31:12.885592 kernel: Write protecting the kernel read-only data: 36864k Apr 14 13:31:12.885598 kernel: Freeing unused kernel image (rodata/data gap) memory: 1824K Apr 14 13:31:12.885603 kernel: Run /init as init process Apr 14 13:31:12.885609 kernel: with arguments: Apr 14 13:31:12.885614 kernel: /init Apr 14 13:31:12.885621 kernel: with environment: Apr 14 13:31:12.885627 kernel: HOME=/ Apr 14 13:31:12.885632 kernel: TERM=linux Apr 14 13:31:12.885640 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 14 13:31:12.885648 systemd[1]: Detected virtualization kvm. Apr 14 13:31:12.885654 systemd[1]: Detected architecture x86-64. Apr 14 13:31:12.885660 systemd[1]: Running in initrd. Apr 14 13:31:12.885665 systemd[1]: No hostname configured, using default hostname. Apr 14 13:31:12.885672 systemd[1]: Hostname set to . Apr 14 13:31:12.885678 systemd[1]: Initializing machine ID from VM UUID. Apr 14 13:31:12.885684 systemd[1]: Queued start job for default target initrd.target. Apr 14 13:31:12.885690 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 14 13:31:12.885696 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 14 13:31:12.885702 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 14 13:31:12.885708 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 14 13:31:12.885733 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 14 13:31:12.885748 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 14 13:31:12.885794 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 14 13:31:12.885801 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 14 13:31:12.885817 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 14 13:31:12.885825 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 14 13:31:12.885831 systemd[1]: Reached target paths.target - Path Units. Apr 14 13:31:12.885837 systemd[1]: Reached target slices.target - Slice Units. Apr 14 13:31:12.885843 systemd[1]: Reached target swap.target - Swaps. Apr 14 13:31:12.885849 systemd[1]: Reached target timers.target - Timer Units. Apr 14 13:31:12.885855 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 14 13:31:12.885861 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 14 13:31:12.885869 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 14 13:31:12.885875 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 14 13:31:12.885882 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 14 13:31:12.885888 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 14 13:31:12.885894 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 14 13:31:12.885900 systemd[1]: Reached target sockets.target - Socket Units. Apr 14 13:31:12.885906 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 14 13:31:12.885912 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 14 13:31:12.885919 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 14 13:31:12.885924 systemd[1]: Starting systemd-fsck-usr.service... Apr 14 13:31:12.885931 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 14 13:31:12.885938 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 14 13:31:12.885944 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 14 13:31:12.885951 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 14 13:31:12.885957 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 14 13:31:12.885962 systemd[1]: Finished systemd-fsck-usr.service. Apr 14 13:31:12.885985 systemd-journald[194]: Collecting audit messages is disabled. Apr 14 13:31:12.886003 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 14 13:31:12.886010 systemd-journald[194]: Journal started Apr 14 13:31:12.886025 systemd-journald[194]: Runtime Journal (/run/log/journal/75e41f6a4960414a90ed3ff16c8d1f56) is 6.0M, max 48.4M, 42.3M free. Apr 14 13:31:12.873083 systemd-modules-load[195]: Inserted module 'overlay' Apr 14 13:31:12.963375 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 14 13:31:12.963397 kernel: Bridge firewalling registered Apr 14 13:31:12.896416 systemd-modules-load[195]: Inserted module 'br_netfilter' Apr 14 13:31:12.971906 systemd[1]: Started systemd-journald.service - Journal Service. Apr 14 13:31:12.972392 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 14 13:31:12.976037 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 14 13:31:12.979463 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 14 13:31:12.990611 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 14 13:31:12.992998 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 14 13:31:12.994971 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 14 13:31:12.998370 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 14 13:31:13.007480 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 14 13:31:13.008945 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 14 13:31:13.010480 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 14 13:31:13.012579 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 14 13:31:13.032844 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 14 13:31:13.038297 systemd-resolved[224]: Positive Trust Anchors: Apr 14 13:31:13.038313 systemd-resolved[224]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 14 13:31:13.038338 systemd-resolved[224]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 14 13:31:13.057144 dracut-cmdline[234]: dracut-dracut-053 Apr 14 13:31:13.057144 dracut-cmdline[234]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c1ba97db2f6278922cfc5bd0ca74b4bb573fca2c3aed19c121a34271e693e156 Apr 14 13:31:13.039532 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 14 13:31:13.040459 systemd-resolved[224]: Defaulting to hostname 'linux'. Apr 14 13:31:13.042173 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 14 13:31:13.043855 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 14 13:31:13.120461 kernel: SCSI subsystem initialized Apr 14 13:31:13.129665 kernel: Loading iSCSI transport class v2.0-870. Apr 14 13:31:13.140700 kernel: iscsi: registered transport (tcp) Apr 14 13:31:13.158677 kernel: iscsi: registered transport (qla4xxx) Apr 14 13:31:13.158756 kernel: QLogic iSCSI HBA Driver Apr 14 13:31:13.191253 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 14 13:31:13.200472 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 14 13:31:13.221555 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 14 13:31:13.221601 kernel: device-mapper: uevent: version 1.0.3 Apr 14 13:31:13.223083 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 14 13:31:13.260440 kernel: raid6: avx512x4 gen() 42953 MB/s Apr 14 13:31:13.277408 kernel: raid6: avx512x2 gen() 41871 MB/s Apr 14 13:31:13.294284 kernel: raid6: avx512x1 gen() 43146 MB/s Apr 14 13:31:13.311479 kernel: raid6: avx2x4 gen() 37491 MB/s Apr 14 13:31:13.328476 kernel: raid6: avx2x2 gen() 36799 MB/s Apr 14 13:31:13.345989 kernel: raid6: avx2x1 gen() 26594 MB/s Apr 14 13:31:13.346160 kernel: raid6: using algorithm avx512x1 gen() 43146 MB/s Apr 14 13:31:13.363925 kernel: raid6: .... xor() 28796 MB/s, rmw enabled Apr 14 13:31:13.364050 kernel: raid6: using avx512x2 recovery algorithm Apr 14 13:31:13.382270 kernel: xor: automatically using best checksumming function avx Apr 14 13:31:13.503454 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 14 13:31:13.513368 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 14 13:31:13.522618 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 14 13:31:13.532803 systemd-udevd[418]: Using default interface naming scheme 'v255'. Apr 14 13:31:13.535502 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 14 13:31:13.538313 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 14 13:31:13.552649 dracut-pre-trigger[428]: rd.md=0: removing MD RAID activation Apr 14 13:31:13.576440 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 14 13:31:13.590706 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 14 13:31:13.623380 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 14 13:31:13.634571 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 14 13:31:13.642693 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 14 13:31:13.644630 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 14 13:31:13.660155 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Apr 14 13:31:13.648512 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 14 13:31:13.650547 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 14 13:31:13.667339 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Apr 14 13:31:13.660476 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 14 13:31:13.672553 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 14 13:31:13.683438 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 14 13:31:13.683481 kernel: GPT:9289727 != 19775487 Apr 14 13:31:13.683489 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 14 13:31:13.683497 kernel: GPT:9289727 != 19775487 Apr 14 13:31:13.683513 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 14 13:31:13.683520 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 14 13:31:13.680321 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 14 13:31:13.694412 kernel: cryptd: max_cpu_qlen set to 1000 Apr 14 13:31:13.685687 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 14 13:31:13.688862 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 14 13:31:13.688966 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 14 13:31:13.691274 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 14 13:31:13.708244 kernel: libata version 3.00 loaded. Apr 14 13:31:13.709505 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 14 13:31:13.716868 kernel: AVX2 version of gcm_enc/dec engaged. Apr 14 13:31:13.716889 kernel: AES CTR mode by8 optimization enabled Apr 14 13:31:13.716897 kernel: ahci 0000:00:1f.2: version 3.0 Apr 14 13:31:13.716963 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 14 13:31:13.729327 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Apr 14 13:31:13.729348 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Apr 14 13:31:13.729478 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Apr 14 13:31:13.742315 kernel: BTRFS: device fsid de1edd48-4571-4695-92f0-7af6e33c4e3d devid 1 transid 31 /dev/vda3 scanned by (udev-worker) (461) Apr 14 13:31:13.742410 kernel: scsi host0: ahci Apr 14 13:31:13.742533 kernel: scsi host1: ahci Apr 14 13:31:13.742605 kernel: scsi host2: ahci Apr 14 13:31:13.742674 kernel: scsi host3: ahci Apr 14 13:31:13.742763 kernel: scsi host4: ahci Apr 14 13:31:13.742829 kernel: scsi host5: ahci Apr 14 13:31:13.742897 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Apr 14 13:31:13.742905 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Apr 14 13:31:13.742912 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Apr 14 13:31:13.742918 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Apr 14 13:31:13.742927 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Apr 14 13:31:13.742934 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Apr 14 13:31:13.742941 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (465) Apr 14 13:31:13.750997 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Apr 14 13:31:13.847579 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Apr 14 13:31:13.847891 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 14 13:31:13.861816 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Apr 14 13:31:13.864289 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Apr 14 13:31:13.872529 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Apr 14 13:31:13.895772 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 14 13:31:13.899059 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 14 13:31:13.909073 disk-uuid[557]: Primary Header is updated. Apr 14 13:31:13.909073 disk-uuid[557]: Secondary Entries is updated. Apr 14 13:31:13.909073 disk-uuid[557]: Secondary Header is updated. Apr 14 13:31:13.915263 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 14 13:31:13.919262 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 14 13:31:13.919526 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 14 13:31:14.051429 kernel: ata5: SATA link down (SStatus 0 SControl 300) Apr 14 13:31:14.051581 kernel: ata4: SATA link down (SStatus 0 SControl 300) Apr 14 13:31:14.054358 kernel: ata6: SATA link down (SStatus 0 SControl 300) Apr 14 13:31:14.054468 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Apr 14 13:31:14.057391 kernel: ata1: SATA link down (SStatus 0 SControl 300) Apr 14 13:31:14.057513 kernel: ata2: SATA link down (SStatus 0 SControl 300) Apr 14 13:31:14.058284 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Apr 14 13:31:14.060156 kernel: ata3.00: applying bridge limits Apr 14 13:31:14.060300 kernel: ata3.00: configured for UDMA/100 Apr 14 13:31:14.062275 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Apr 14 13:31:14.124992 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Apr 14 13:31:14.125467 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 14 13:31:14.147310 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Apr 14 13:31:14.921008 disk-uuid[559]: The operation has completed successfully. Apr 14 13:31:14.923447 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 14 13:31:14.949244 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 14 13:31:14.949338 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 14 13:31:14.955469 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 14 13:31:14.960275 sh[594]: Success Apr 14 13:31:14.971273 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Apr 14 13:31:15.001901 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 14 13:31:15.010562 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 14 13:31:15.012420 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 14 13:31:15.028318 kernel: BTRFS info (device dm-0): first mount of filesystem de1edd48-4571-4695-92f0-7af6e33c4e3d Apr 14 13:31:15.028363 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Apr 14 13:31:15.028374 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 14 13:31:15.032405 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 14 13:31:15.032512 kernel: BTRFS info (device dm-0): using free space tree Apr 14 13:31:15.038848 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 14 13:31:15.039386 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 14 13:31:15.055447 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 14 13:31:15.057759 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 14 13:31:15.070438 kernel: BTRFS info (device vda6): first mount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 14 13:31:15.070483 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 14 13:31:15.070492 kernel: BTRFS info (device vda6): using free space tree Apr 14 13:31:15.074250 kernel: BTRFS info (device vda6): auto enabling async discard Apr 14 13:31:15.081272 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 14 13:31:15.083984 kernel: BTRFS info (device vda6): last unmount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 14 13:31:15.089316 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 14 13:31:15.095414 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 14 13:31:15.148349 ignition[697]: Ignition 2.19.0 Apr 14 13:31:15.148371 ignition[697]: Stage: fetch-offline Apr 14 13:31:15.148404 ignition[697]: no configs at "/usr/lib/ignition/base.d" Apr 14 13:31:15.148413 ignition[697]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 14 13:31:15.148526 ignition[697]: parsed url from cmdline: "" Apr 14 13:31:15.148530 ignition[697]: no config URL provided Apr 14 13:31:15.148536 ignition[697]: reading system config file "/usr/lib/ignition/user.ign" Apr 14 13:31:15.148544 ignition[697]: no config at "/usr/lib/ignition/user.ign" Apr 14 13:31:15.148571 ignition[697]: op(1): [started] loading QEMU firmware config module Apr 14 13:31:15.148576 ignition[697]: op(1): executing: "modprobe" "qemu_fw_cfg" Apr 14 13:31:15.158122 ignition[697]: op(1): [finished] loading QEMU firmware config module Apr 14 13:31:15.165788 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 14 13:31:15.178823 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 14 13:31:15.196343 systemd-networkd[784]: lo: Link UP Apr 14 13:31:15.196361 systemd-networkd[784]: lo: Gained carrier Apr 14 13:31:15.197410 systemd-networkd[784]: Enumeration completed Apr 14 13:31:15.197637 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 14 13:31:15.198092 systemd[1]: Reached target network.target - Network. Apr 14 13:31:15.198290 systemd-networkd[784]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 14 13:31:15.198293 systemd-networkd[784]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 14 13:31:15.199128 systemd-networkd[784]: eth0: Link UP Apr 14 13:31:15.199130 systemd-networkd[784]: eth0: Gained carrier Apr 14 13:31:15.199136 systemd-networkd[784]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 14 13:31:15.255543 systemd-networkd[784]: eth0: DHCPv4 address 10.0.0.8/16, gateway 10.0.0.1 acquired from 10.0.0.1 Apr 14 13:31:15.279357 ignition[697]: parsing config with SHA512: cdac23f2448b8a6ea7d8a6432cd3f437dd9c2b81b2043b818fe12f4ea50cea7ff7af85d0c3424c91896f938218b453725c72e65e71dc0a8edaa87b5bd464a30b Apr 14 13:31:15.286124 unknown[697]: fetched base config from "system" Apr 14 13:31:15.286136 unknown[697]: fetched user config from "qemu" Apr 14 13:31:15.286656 ignition[697]: fetch-offline: fetch-offline passed Apr 14 13:31:15.286758 ignition[697]: Ignition finished successfully Apr 14 13:31:15.294593 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 14 13:31:15.299884 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Apr 14 13:31:15.315717 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 14 13:31:15.332800 ignition[788]: Ignition 2.19.0 Apr 14 13:31:15.332818 ignition[788]: Stage: kargs Apr 14 13:31:15.332954 ignition[788]: no configs at "/usr/lib/ignition/base.d" Apr 14 13:31:15.332961 ignition[788]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 14 13:31:15.333793 ignition[788]: kargs: kargs passed Apr 14 13:31:15.333835 ignition[788]: Ignition finished successfully Apr 14 13:31:15.345624 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 14 13:31:15.355540 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 14 13:31:15.366464 ignition[797]: Ignition 2.19.0 Apr 14 13:31:15.366480 ignition[797]: Stage: disks Apr 14 13:31:15.366634 ignition[797]: no configs at "/usr/lib/ignition/base.d" Apr 14 13:31:15.366642 ignition[797]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 14 13:31:15.369787 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 14 13:31:15.367385 ignition[797]: disks: disks passed Apr 14 13:31:15.374490 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 14 13:31:15.367425 ignition[797]: Ignition finished successfully Apr 14 13:31:15.377672 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 14 13:31:15.384508 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 14 13:31:15.386347 systemd[1]: Reached target sysinit.target - System Initialization. Apr 14 13:31:15.388020 systemd[1]: Reached target basic.target - Basic System. Apr 14 13:31:15.403806 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 14 13:31:15.415634 systemd-resolved[224]: Detected conflict on linux IN A 10.0.0.8 Apr 14 13:31:15.415657 systemd-resolved[224]: Hostname conflict, changing published hostname from 'linux' to 'linux10'. Apr 14 13:31:15.418206 systemd-fsck[808]: ROOT: clean, 14/553520 files, 52654/553472 blocks Apr 14 13:31:15.422212 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 14 13:31:15.423086 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 14 13:31:15.502259 kernel: EXT4-fs (vda9): mounted filesystem e02793bf-3e0d-4c7e-b11a-92c664da7ce3 r/w with ordered data mode. Quota mode: none. Apr 14 13:31:15.502860 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 14 13:31:15.503378 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 14 13:31:15.528803 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 14 13:31:15.531479 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 14 13:31:15.534110 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Apr 14 13:31:15.540919 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (816) Apr 14 13:31:15.540944 kernel: BTRFS info (device vda6): first mount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 14 13:31:15.534155 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 14 13:31:15.546550 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 14 13:31:15.546573 kernel: BTRFS info (device vda6): using free space tree Apr 14 13:31:15.534179 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 14 13:31:15.539688 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 14 13:31:15.545069 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 14 13:31:15.553277 kernel: BTRFS info (device vda6): auto enabling async discard Apr 14 13:31:15.554085 systemd-resolved[224]: Detected conflict on linux10 IN A 10.0.0.8 Apr 14 13:31:15.554124 systemd-resolved[224]: Hostname conflict, changing published hostname from 'linux10' to 'linux12'. Apr 14 13:31:15.563697 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 14 13:31:15.584437 initrd-setup-root[840]: cut: /sysroot/etc/passwd: No such file or directory Apr 14 13:31:15.590178 initrd-setup-root[847]: cut: /sysroot/etc/group: No such file or directory Apr 14 13:31:15.595124 initrd-setup-root[854]: cut: /sysroot/etc/shadow: No such file or directory Apr 14 13:31:15.599847 initrd-setup-root[861]: cut: /sysroot/etc/gshadow: No such file or directory Apr 14 13:31:15.667388 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 14 13:31:15.673412 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 14 13:31:15.674602 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 14 13:31:15.686561 kernel: BTRFS info (device vda6): last unmount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 14 13:31:15.701403 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 14 13:31:15.711753 ignition[931]: INFO : Ignition 2.19.0 Apr 14 13:31:15.711753 ignition[931]: INFO : Stage: mount Apr 14 13:31:15.714882 ignition[931]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 14 13:31:15.714882 ignition[931]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 14 13:31:15.714882 ignition[931]: INFO : mount: mount passed Apr 14 13:31:15.714882 ignition[931]: INFO : Ignition finished successfully Apr 14 13:31:15.723271 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 14 13:31:15.734354 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 14 13:31:16.025727 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 14 13:31:16.033720 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 14 13:31:16.042632 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (943) Apr 14 13:31:16.042672 kernel: BTRFS info (device vda6): first mount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 14 13:31:16.042681 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 14 13:31:16.045375 kernel: BTRFS info (device vda6): using free space tree Apr 14 13:31:16.049363 kernel: BTRFS info (device vda6): auto enabling async discard Apr 14 13:31:16.050156 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 14 13:31:16.074921 ignition[960]: INFO : Ignition 2.19.0 Apr 14 13:31:16.074921 ignition[960]: INFO : Stage: files Apr 14 13:31:16.074921 ignition[960]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 14 13:31:16.074921 ignition[960]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 14 13:31:16.082113 ignition[960]: DEBUG : files: compiled without relabeling support, skipping Apr 14 13:31:16.082113 ignition[960]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 14 13:31:16.082113 ignition[960]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 14 13:31:16.082113 ignition[960]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 14 13:31:16.082113 ignition[960]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 14 13:31:16.082113 ignition[960]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 14 13:31:16.082113 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 14 13:31:16.082113 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Apr 14 13:31:16.080193 unknown[960]: wrote ssh authorized keys file for user: core Apr 14 13:31:16.107625 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 14 13:31:16.193672 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 14 13:31:16.193672 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Apr 14 13:31:16.199658 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.35.1-x86-64.raw: attempt #1 Apr 14 13:31:16.282613 systemd-networkd[784]: eth0: Gained IPv6LL Apr 14 13:31:16.556442 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 14 13:31:17.426725 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Apr 14 13:31:17.426725 ignition[960]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Apr 14 13:31:17.432359 ignition[960]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 14 13:31:17.435179 ignition[960]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 14 13:31:17.435179 ignition[960]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Apr 14 13:31:17.435179 ignition[960]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Apr 14 13:31:17.441717 ignition[960]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Apr 14 13:31:17.441717 ignition[960]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Apr 14 13:31:17.441717 ignition[960]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Apr 14 13:31:17.441717 ignition[960]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Apr 14 13:31:17.461890 ignition[960]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Apr 14 13:31:17.467053 ignition[960]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Apr 14 13:31:17.469448 ignition[960]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Apr 14 13:31:17.469448 ignition[960]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Apr 14 13:31:17.469448 ignition[960]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Apr 14 13:31:17.469448 ignition[960]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 14 13:31:17.469448 ignition[960]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 14 13:31:17.469448 ignition[960]: INFO : files: files passed Apr 14 13:31:17.469448 ignition[960]: INFO : Ignition finished successfully Apr 14 13:31:17.472372 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 14 13:31:17.488467 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 14 13:31:17.490915 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 14 13:31:17.496874 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 14 13:31:17.497006 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 14 13:31:17.505639 initrd-setup-root-after-ignition[988]: grep: /sysroot/oem/oem-release: No such file or directory Apr 14 13:31:17.511555 initrd-setup-root-after-ignition[990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 14 13:31:17.514281 initrd-setup-root-after-ignition[990]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 14 13:31:17.517808 initrd-setup-root-after-ignition[994]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 14 13:31:17.518875 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 14 13:31:17.523459 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 14 13:31:17.533438 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 14 13:31:17.560790 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 14 13:31:17.560888 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 14 13:31:17.564370 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 14 13:31:17.567391 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 14 13:31:17.568913 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 14 13:31:17.583683 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 14 13:31:17.595056 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 14 13:31:17.614731 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 14 13:31:17.624604 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 14 13:31:17.627342 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 14 13:31:17.630961 systemd[1]: Stopped target timers.target - Timer Units. Apr 14 13:31:17.634374 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 14 13:31:17.634490 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 14 13:31:17.639696 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 14 13:31:17.643098 systemd[1]: Stopped target basic.target - Basic System. Apr 14 13:31:17.644580 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 14 13:31:17.650768 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 14 13:31:17.654801 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 14 13:31:17.656509 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 14 13:31:17.658139 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 14 13:31:17.662527 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 14 13:31:17.664269 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 14 13:31:17.668278 systemd[1]: Stopped target swap.target - Swaps. Apr 14 13:31:17.670715 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 14 13:31:17.670844 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 14 13:31:17.674723 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 14 13:31:17.677679 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 14 13:31:17.680940 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 14 13:31:17.681052 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 14 13:31:17.682680 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 14 13:31:17.682793 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 14 13:31:17.689140 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 14 13:31:17.689264 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 14 13:31:17.692414 systemd[1]: Stopped target paths.target - Path Units. Apr 14 13:31:17.695094 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 14 13:31:17.698773 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 14 13:31:17.699328 systemd[1]: Stopped target slices.target - Slice Units. Apr 14 13:31:17.702813 systemd[1]: Stopped target sockets.target - Socket Units. Apr 14 13:31:17.705989 systemd[1]: iscsid.socket: Deactivated successfully. Apr 14 13:31:17.706095 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 14 13:31:17.708458 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 14 13:31:17.708553 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 14 13:31:17.711199 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 14 13:31:17.711796 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 14 13:31:17.716560 systemd[1]: ignition-files.service: Deactivated successfully. Apr 14 13:31:17.716636 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 14 13:31:17.727425 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 14 13:31:17.727501 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 14 13:31:17.727585 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 14 13:31:17.730848 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 14 13:31:17.734779 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 14 13:31:17.734885 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 14 13:31:17.740057 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 14 13:31:17.740148 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 14 13:31:17.740763 ignition[1014]: INFO : Ignition 2.19.0 Apr 14 13:31:17.740763 ignition[1014]: INFO : Stage: umount Apr 14 13:31:17.740763 ignition[1014]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 14 13:31:17.740763 ignition[1014]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 14 13:31:17.741930 ignition[1014]: INFO : umount: umount passed Apr 14 13:31:17.741930 ignition[1014]: INFO : Ignition finished successfully Apr 14 13:31:17.743167 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 14 13:31:17.743279 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 14 13:31:17.743906 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 14 13:31:17.743973 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 14 13:31:17.745977 systemd[1]: Stopped target network.target - Network. Apr 14 13:31:17.746445 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 14 13:31:17.746490 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 14 13:31:17.746680 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 14 13:31:17.746701 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 14 13:31:17.746909 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 14 13:31:17.746930 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 14 13:31:17.747145 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 14 13:31:17.747169 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 14 13:31:17.747764 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 14 13:31:17.748151 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 14 13:31:17.782318 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 14 13:31:17.782442 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 14 13:31:17.784772 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 14 13:31:17.784810 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 14 13:31:17.790327 systemd-networkd[784]: eth0: DHCPv6 lease lost Apr 14 13:31:17.792872 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 14 13:31:17.794997 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 14 13:31:17.798116 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 14 13:31:17.798165 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 14 13:31:17.806347 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 14 13:31:17.807775 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 14 13:31:17.807825 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 14 13:31:17.816888 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 14 13:31:17.816963 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 14 13:31:17.819995 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 14 13:31:17.820050 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 14 13:31:17.823417 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 14 13:31:17.828964 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 14 13:31:17.836631 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 14 13:31:17.836758 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 14 13:31:17.839006 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 14 13:31:17.839119 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 14 13:31:17.842627 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 14 13:31:17.842654 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 14 13:31:17.845405 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 14 13:31:17.845428 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 14 13:31:17.847119 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 14 13:31:17.847154 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 14 13:31:17.850348 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 14 13:31:17.850382 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 14 13:31:17.854394 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 14 13:31:17.854426 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 14 13:31:17.870602 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 14 13:31:17.870703 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 14 13:31:17.870783 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 14 13:31:17.878250 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 14 13:31:17.878292 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 14 13:31:17.880270 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 14 13:31:17.880345 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 14 13:31:17.889021 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 14 13:31:17.889123 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 14 13:31:17.891005 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 14 13:31:17.894499 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 14 13:31:17.894551 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 14 13:31:17.902418 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 14 13:31:17.911398 systemd[1]: Switching root. Apr 14 13:31:17.938178 systemd-journald[194]: Journal stopped Apr 14 13:31:18.720331 systemd-journald[194]: Received SIGTERM from PID 1 (systemd). Apr 14 13:31:18.720405 kernel: SELinux: policy capability network_peer_controls=1 Apr 14 13:31:18.720425 kernel: SELinux: policy capability open_perms=1 Apr 14 13:31:18.720439 kernel: SELinux: policy capability extended_socket_class=1 Apr 14 13:31:18.720451 kernel: SELinux: policy capability always_check_network=0 Apr 14 13:31:18.720464 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 14 13:31:18.720476 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 14 13:31:18.720490 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 14 13:31:18.720506 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 14 13:31:18.720514 kernel: audit: type=1403 audit(1776173478.055:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 14 13:31:18.720528 systemd[1]: Successfully loaded SELinux policy in 34.441ms. Apr 14 13:31:18.720551 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 8.603ms. Apr 14 13:31:18.720567 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 14 13:31:18.720582 systemd[1]: Detected virtualization kvm. Apr 14 13:31:18.720596 systemd[1]: Detected architecture x86-64. Apr 14 13:31:18.720612 systemd[1]: Detected first boot. Apr 14 13:31:18.720631 systemd[1]: Initializing machine ID from VM UUID. Apr 14 13:31:18.720646 zram_generator::config[1059]: No configuration found. Apr 14 13:31:18.720661 systemd[1]: Populated /etc with preset unit settings. Apr 14 13:31:18.720675 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 14 13:31:18.720687 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 14 13:31:18.720701 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 14 13:31:18.720717 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 14 13:31:18.720731 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 14 13:31:18.720817 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 14 13:31:18.720837 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 14 13:31:18.720849 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 14 13:31:18.720862 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 14 13:31:18.720876 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 14 13:31:18.720890 systemd[1]: Created slice user.slice - User and Session Slice. Apr 14 13:31:18.720902 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 14 13:31:18.720915 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 14 13:31:18.720928 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 14 13:31:18.720941 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 14 13:31:18.720955 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 14 13:31:18.720968 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 14 13:31:18.720985 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Apr 14 13:31:18.720999 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 14 13:31:18.721013 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 14 13:31:18.721026 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 14 13:31:18.721039 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 14 13:31:18.721054 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 14 13:31:18.721067 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 14 13:31:18.721081 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 14 13:31:18.721094 systemd[1]: Reached target slices.target - Slice Units. Apr 14 13:31:18.721106 systemd[1]: Reached target swap.target - Swaps. Apr 14 13:31:18.721118 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 14 13:31:18.721129 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 14 13:31:18.721141 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 14 13:31:18.721154 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 14 13:31:18.721168 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 14 13:31:18.721181 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 14 13:31:18.721193 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 14 13:31:18.721205 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 14 13:31:18.721263 systemd[1]: Mounting media.mount - External Media Directory... Apr 14 13:31:18.721281 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 14 13:31:18.721296 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 14 13:31:18.721311 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 14 13:31:18.721324 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 14 13:31:18.721341 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 14 13:31:18.721355 systemd[1]: Reached target machines.target - Containers. Apr 14 13:31:18.721373 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 14 13:31:18.721386 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 14 13:31:18.721400 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 14 13:31:18.721413 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 14 13:31:18.721426 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 14 13:31:18.721438 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 14 13:31:18.721454 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 14 13:31:18.721466 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 14 13:31:18.721478 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 14 13:31:18.721491 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 14 13:31:18.721503 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 14 13:31:18.721515 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 14 13:31:18.721527 kernel: fuse: init (API version 7.39) Apr 14 13:31:18.721540 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 14 13:31:18.721552 systemd[1]: Stopped systemd-fsck-usr.service. Apr 14 13:31:18.721568 kernel: loop: module loaded Apr 14 13:31:18.721580 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 14 13:31:18.721595 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 14 13:31:18.721608 kernel: ACPI: bus type drm_connector registered Apr 14 13:31:18.721621 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 14 13:31:18.721633 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 14 13:31:18.721670 systemd-journald[1143]: Collecting audit messages is disabled. Apr 14 13:31:18.721693 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 14 13:31:18.721709 systemd-journald[1143]: Journal started Apr 14 13:31:18.721735 systemd-journald[1143]: Runtime Journal (/run/log/journal/75e41f6a4960414a90ed3ff16c8d1f56) is 6.0M, max 48.4M, 42.3M free. Apr 14 13:31:18.424837 systemd[1]: Queued start job for default target multi-user.target. Apr 14 13:31:18.456270 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Apr 14 13:31:18.456664 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 14 13:31:18.726916 systemd[1]: verity-setup.service: Deactivated successfully. Apr 14 13:31:18.726953 systemd[1]: Stopped verity-setup.service. Apr 14 13:31:18.728070 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 14 13:31:18.734915 systemd[1]: Started systemd-journald.service - Journal Service. Apr 14 13:31:18.735940 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 14 13:31:18.737643 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 14 13:31:18.739412 systemd[1]: Mounted media.mount - External Media Directory. Apr 14 13:31:18.741097 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 14 13:31:18.742848 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 14 13:31:18.744564 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 14 13:31:18.746309 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 14 13:31:18.748513 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 14 13:31:18.750721 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 14 13:31:18.750944 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 14 13:31:18.753203 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 14 13:31:18.753483 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 14 13:31:18.755470 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 14 13:31:18.755644 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 14 13:31:18.757502 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 14 13:31:18.757673 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 14 13:31:18.759737 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 14 13:31:18.759942 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 14 13:31:18.761785 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 14 13:31:18.762257 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 14 13:31:18.764101 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 14 13:31:18.766662 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 14 13:31:18.768962 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 14 13:31:18.778518 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 14 13:31:18.783958 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 14 13:31:18.790348 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 14 13:31:18.793268 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 14 13:31:18.795258 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 14 13:31:18.795289 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 14 13:31:18.798473 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Apr 14 13:31:18.806453 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 14 13:31:18.809489 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 14 13:31:18.811245 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 14 13:31:18.812841 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 14 13:31:18.815330 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 14 13:31:18.817594 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 14 13:31:18.821541 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 14 13:31:18.821647 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 14 13:31:18.823585 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 14 13:31:18.828546 systemd-journald[1143]: Time spent on flushing to /var/log/journal/75e41f6a4960414a90ed3ff16c8d1f56 is 16.754ms for 953 entries. Apr 14 13:31:18.828546 systemd-journald[1143]: System Journal (/var/log/journal/75e41f6a4960414a90ed3ff16c8d1f56) is 8.0M, max 195.6M, 187.6M free. Apr 14 13:31:18.855354 systemd-journald[1143]: Received client request to flush runtime journal. Apr 14 13:31:18.855400 kernel: loop0: detected capacity change from 0 to 140768 Apr 14 13:31:18.830518 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 14 13:31:18.839971 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 14 13:31:18.845379 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 14 13:31:18.852625 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 14 13:31:18.863360 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 14 13:31:18.865903 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 14 13:31:18.869025 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 14 13:31:18.871254 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 14 13:31:18.873735 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 14 13:31:18.880692 udevadm[1177]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Apr 14 13:31:18.881480 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 14 13:31:18.885282 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 14 13:31:18.894086 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Apr 14 13:31:18.896456 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 14 13:31:18.901392 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 14 13:31:18.912625 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 14 13:31:18.915299 kernel: loop1: detected capacity change from 0 to 142488 Apr 14 13:31:18.915695 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Apr 14 13:31:18.938480 systemd-tmpfiles[1191]: ACLs are not supported, ignoring. Apr 14 13:31:18.938495 systemd-tmpfiles[1191]: ACLs are not supported, ignoring. Apr 14 13:31:18.944947 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 14 13:31:18.958406 kernel: loop2: detected capacity change from 0 to 217752 Apr 14 13:31:18.988242 kernel: loop3: detected capacity change from 0 to 140768 Apr 14 13:31:19.001330 kernel: loop4: detected capacity change from 0 to 142488 Apr 14 13:31:19.012289 kernel: loop5: detected capacity change from 0 to 217752 Apr 14 13:31:19.022897 (sd-merge)[1197]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Apr 14 13:31:19.024017 (sd-merge)[1197]: Merged extensions into '/usr'. Apr 14 13:31:19.028990 systemd[1]: Reloading requested from client PID 1174 ('systemd-sysext') (unit systemd-sysext.service)... Apr 14 13:31:19.029016 systemd[1]: Reloading... Apr 14 13:31:19.076268 zram_generator::config[1219]: No configuration found. Apr 14 13:31:19.143137 ldconfig[1169]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 14 13:31:19.185129 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 14 13:31:19.216299 systemd[1]: Reloading finished in 186 ms. Apr 14 13:31:19.248950 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 14 13:31:19.252199 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 14 13:31:19.268535 systemd[1]: Starting ensure-sysext.service... Apr 14 13:31:19.271051 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 14 13:31:19.275309 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 14 13:31:19.279410 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 14 13:31:19.283183 systemd[1]: Reloading requested from client PID 1260 ('systemctl') (unit ensure-sysext.service)... Apr 14 13:31:19.283205 systemd[1]: Reloading... Apr 14 13:31:19.292987 systemd-tmpfiles[1261]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 14 13:31:19.293358 systemd-tmpfiles[1261]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 14 13:31:19.294042 systemd-tmpfiles[1261]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 14 13:31:19.294359 systemd-tmpfiles[1261]: ACLs are not supported, ignoring. Apr 14 13:31:19.294420 systemd-tmpfiles[1261]: ACLs are not supported, ignoring. Apr 14 13:31:19.297190 systemd-tmpfiles[1261]: Detected autofs mount point /boot during canonicalization of boot. Apr 14 13:31:19.297326 systemd-tmpfiles[1261]: Skipping /boot Apr 14 13:31:19.302741 systemd-udevd[1263]: Using default interface naming scheme 'v255'. Apr 14 13:31:19.303605 systemd-tmpfiles[1261]: Detected autofs mount point /boot during canonicalization of boot. Apr 14 13:31:19.303643 systemd-tmpfiles[1261]: Skipping /boot Apr 14 13:31:19.330286 zram_generator::config[1288]: No configuration found. Apr 14 13:31:19.361561 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (1293) Apr 14 13:31:19.402305 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Apr 14 13:31:19.410273 kernel: ACPI: button: Power Button [PWRF] Apr 14 13:31:19.431466 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Apr 14 13:31:19.431699 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Apr 14 13:31:19.431853 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Apr 14 13:31:19.434361 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Apr 14 13:31:19.434638 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 14 13:31:19.466256 kernel: mousedev: PS/2 mouse device common for all mice Apr 14 13:31:19.479838 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Apr 14 13:31:19.480451 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Apr 14 13:31:19.484113 systemd[1]: Reloading finished in 200 ms. Apr 14 13:31:19.542438 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 14 13:31:19.551661 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 14 13:31:19.588088 systemd[1]: Finished ensure-sysext.service. Apr 14 13:31:19.605673 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 14 13:31:19.612414 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 14 13:31:19.615309 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 14 13:31:19.617171 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 14 13:31:19.618404 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 14 13:31:19.623515 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 14 13:31:19.626445 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 14 13:31:19.629698 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 14 13:31:19.631600 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 14 13:31:19.634029 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 14 13:31:19.637372 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 14 13:31:19.639705 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 14 13:31:19.646596 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 14 13:31:19.650479 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 14 13:31:19.654105 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 14 13:31:19.657416 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 14 13:31:19.659521 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 14 13:31:19.660793 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Apr 14 13:31:19.663166 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 14 13:31:19.663515 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 14 13:31:19.665662 augenrules[1386]: No rules Apr 14 13:31:19.665898 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 14 13:31:19.666031 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 14 13:31:19.670056 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 14 13:31:19.672269 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 14 13:31:19.672369 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 14 13:31:19.674815 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 14 13:31:19.674932 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 14 13:31:19.677124 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 14 13:31:19.677789 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 14 13:31:19.695554 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Apr 14 13:31:19.695691 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 14 13:31:19.695776 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 14 13:31:19.697732 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 14 13:31:19.703726 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 14 13:31:19.704495 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 14 13:31:19.704892 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 14 13:31:19.708247 lvm[1402]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 14 13:31:19.707651 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 14 13:31:19.717837 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 14 13:31:19.730518 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Apr 14 13:31:19.778781 systemd-networkd[1377]: lo: Link UP Apr 14 13:31:19.778797 systemd-networkd[1377]: lo: Gained carrier Apr 14 13:31:19.779816 systemd-networkd[1377]: Enumeration completed Apr 14 13:31:19.780386 systemd-networkd[1377]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 14 13:31:19.780397 systemd-networkd[1377]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 14 13:31:19.780973 systemd-networkd[1377]: eth0: Link UP Apr 14 13:31:19.780979 systemd-networkd[1377]: eth0: Gained carrier Apr 14 13:31:19.780988 systemd-networkd[1377]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 14 13:31:19.791597 systemd-resolved[1379]: Positive Trust Anchors: Apr 14 13:31:19.791624 systemd-resolved[1379]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 14 13:31:19.791648 systemd-resolved[1379]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 14 13:31:19.794798 systemd-resolved[1379]: Defaulting to hostname 'linux'. Apr 14 13:31:19.809447 systemd-networkd[1377]: eth0: DHCPv4 address 10.0.0.8/16, gateway 10.0.0.1 acquired from 10.0.0.1 Apr 14 13:31:19.810258 systemd-timesyncd[1383]: Network configuration changed, trying to establish connection. Apr 14 13:31:19.812023 systemd-timesyncd[1383]: Contacted time server 10.0.0.1:123 (10.0.0.1). Apr 14 13:31:19.812071 systemd-timesyncd[1383]: Initial clock synchronization to Tue 2026-04-14 13:31:20.201687 UTC. Apr 14 13:31:19.824545 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 14 13:31:19.828500 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 14 13:31:19.830361 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 14 13:31:19.832371 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 14 13:31:19.834528 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 14 13:31:19.837308 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 14 13:31:19.838921 systemd[1]: Reached target network.target - Network. Apr 14 13:31:19.840324 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 14 13:31:19.842043 systemd[1]: Reached target sysinit.target - System Initialization. Apr 14 13:31:19.844722 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 14 13:31:19.846622 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 14 13:31:19.848450 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 14 13:31:19.850399 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 14 13:31:19.850435 systemd[1]: Reached target paths.target - Path Units. Apr 14 13:31:19.851825 systemd[1]: Reached target time-set.target - System Time Set. Apr 14 13:31:19.853525 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 14 13:31:19.855299 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 14 13:31:19.857182 systemd[1]: Reached target timers.target - Timer Units. Apr 14 13:31:19.859356 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 14 13:31:19.862375 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 14 13:31:19.870130 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 14 13:31:19.872779 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Apr 14 13:31:19.875521 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 14 13:31:19.877805 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 14 13:31:19.879544 systemd[1]: Reached target sockets.target - Socket Units. Apr 14 13:31:19.881046 systemd[1]: Reached target basic.target - Basic System. Apr 14 13:31:19.882433 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 14 13:31:19.882481 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 14 13:31:19.883306 lvm[1423]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 14 13:31:19.883659 systemd[1]: Starting containerd.service - containerd container runtime... Apr 14 13:31:19.886110 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 14 13:31:19.888655 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 14 13:31:19.893405 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 14 13:31:19.895282 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 14 13:31:19.901883 jq[1427]: false Apr 14 13:31:19.901314 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 14 13:31:19.909451 extend-filesystems[1428]: Found loop3 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found loop4 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found loop5 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found sr0 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found vda Apr 14 13:31:19.909451 extend-filesystems[1428]: Found vda1 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found vda2 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found vda3 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found usr Apr 14 13:31:19.909451 extend-filesystems[1428]: Found vda4 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found vda6 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found vda7 Apr 14 13:31:19.909451 extend-filesystems[1428]: Found vda9 Apr 14 13:31:19.909451 extend-filesystems[1428]: Checking size of /dev/vda9 Apr 14 13:31:19.912420 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 14 13:31:19.918528 dbus-daemon[1426]: [system] SELinux support is enabled Apr 14 13:31:19.916416 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 14 13:31:19.919722 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 14 13:31:19.926794 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 14 13:31:19.930067 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 14 13:31:19.930583 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 14 13:31:19.934506 systemd[1]: Starting update-engine.service - Update Engine... Apr 14 13:31:19.939439 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 14 13:31:19.942952 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 14 13:31:19.943400 extend-filesystems[1428]: Resized partition /dev/vda9 Apr 14 13:31:19.948177 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Apr 14 13:31:19.948831 jq[1447]: true Apr 14 13:31:19.950296 extend-filesystems[1448]: resize2fs 1.47.1 (20-May-2024) Apr 14 13:31:19.953725 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (1297) Apr 14 13:31:19.964574 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Apr 14 13:31:19.963618 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 14 13:31:19.963866 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 14 13:31:19.964165 systemd[1]: motdgen.service: Deactivated successfully. Apr 14 13:31:19.964445 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 14 13:31:19.968558 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 14 13:31:19.968743 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 14 13:31:19.971353 update_engine[1445]: I20260414 13:31:19.969787 1445 main.cc:92] Flatcar Update Engine starting Apr 14 13:31:19.974063 update_engine[1445]: I20260414 13:31:19.974010 1445 update_check_scheduler.cc:74] Next update check in 3m38s Apr 14 13:31:19.980320 jq[1453]: true Apr 14 13:31:19.979251 systemd-logind[1442]: Watching system buttons on /dev/input/event1 (Power Button) Apr 14 13:31:19.979267 systemd-logind[1442]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Apr 14 13:31:19.980791 systemd-logind[1442]: New seat seat0. Apr 14 13:31:19.988911 systemd[1]: Started systemd-logind.service - User Login Management. Apr 14 13:31:19.990108 (ntainerd)[1454]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 14 13:31:19.995149 dbus-daemon[1426]: [system] Successfully activated service 'org.freedesktop.systemd1' Apr 14 13:31:19.999624 systemd[1]: Started update-engine.service - Update Engine. Apr 14 13:31:20.001853 tar[1452]: linux-amd64/LICENSE Apr 14 13:31:20.002129 tar[1452]: linux-amd64/helm Apr 14 13:31:20.002661 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 14 13:31:20.002779 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 14 13:31:20.011055 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 14 13:31:20.011204 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 14 13:31:20.045494 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 14 13:31:20.087398 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Apr 14 13:31:20.106134 locksmithd[1479]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 14 13:31:20.110513 extend-filesystems[1448]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Apr 14 13:31:20.110513 extend-filesystems[1448]: old_desc_blocks = 1, new_desc_blocks = 1 Apr 14 13:31:20.110513 extend-filesystems[1448]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Apr 14 13:31:20.117321 sshd_keygen[1449]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 14 13:31:20.110356 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 14 13:31:20.117473 extend-filesystems[1428]: Resized filesystem in /dev/vda9 Apr 14 13:31:20.117581 bash[1480]: Updated "/home/core/.ssh/authorized_keys" Apr 14 13:31:20.110510 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 14 13:31:20.121031 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 14 13:31:20.126015 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Apr 14 13:31:20.140771 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 14 13:31:20.149824 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 14 13:31:20.158914 systemd[1]: issuegen.service: Deactivated successfully. Apr 14 13:31:20.159103 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 14 13:31:20.166609 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 14 13:31:20.180575 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 14 13:31:20.191632 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 14 13:31:20.194422 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Apr 14 13:31:20.196371 systemd[1]: Reached target getty.target - Login Prompts. Apr 14 13:31:20.244307 containerd[1454]: time="2026-04-14T13:31:20.242151717Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Apr 14 13:31:20.263405 containerd[1454]: time="2026-04-14T13:31:20.263144966Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 14 13:31:20.265981 containerd[1454]: time="2026-04-14T13:31:20.265938534Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 14 13:31:20.266077 containerd[1454]: time="2026-04-14T13:31:20.266068532Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 14 13:31:20.266165 containerd[1454]: time="2026-04-14T13:31:20.266156111Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 14 13:31:20.266366 containerd[1454]: time="2026-04-14T13:31:20.266354087Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Apr 14 13:31:20.266407 containerd[1454]: time="2026-04-14T13:31:20.266400456Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Apr 14 13:31:20.266476 containerd[1454]: time="2026-04-14T13:31:20.266466375Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Apr 14 13:31:20.266505 containerd[1454]: time="2026-04-14T13:31:20.266498941Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 14 13:31:20.266724 containerd[1454]: time="2026-04-14T13:31:20.266711476Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 14 13:31:20.266758 containerd[1454]: time="2026-04-14T13:31:20.266752045Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 14 13:31:20.266789 containerd[1454]: time="2026-04-14T13:31:20.266781626Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Apr 14 13:31:20.266813 containerd[1454]: time="2026-04-14T13:31:20.266807600Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 14 13:31:20.266906 containerd[1454]: time="2026-04-14T13:31:20.266897358Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 14 13:31:20.267085 containerd[1454]: time="2026-04-14T13:31:20.267074753Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 14 13:31:20.267728 containerd[1454]: time="2026-04-14T13:31:20.267660029Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 14 13:31:20.267818 containerd[1454]: time="2026-04-14T13:31:20.267810138Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 14 13:31:20.267911 containerd[1454]: time="2026-04-14T13:31:20.267903199Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 14 13:31:20.267966 containerd[1454]: time="2026-04-14T13:31:20.267959160Z" level=info msg="metadata content store policy set" policy=shared Apr 14 13:31:20.275213 containerd[1454]: time="2026-04-14T13:31:20.275174689Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 14 13:31:20.275395 containerd[1454]: time="2026-04-14T13:31:20.275379747Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 14 13:31:20.275456 containerd[1454]: time="2026-04-14T13:31:20.275445654Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Apr 14 13:31:20.275513 containerd[1454]: time="2026-04-14T13:31:20.275502008Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Apr 14 13:31:20.275639 containerd[1454]: time="2026-04-14T13:31:20.275629312Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 14 13:31:20.275859 containerd[1454]: time="2026-04-14T13:31:20.275848753Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 14 13:31:20.276186 containerd[1454]: time="2026-04-14T13:31:20.276153002Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 14 13:31:20.276373 containerd[1454]: time="2026-04-14T13:31:20.276330074Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Apr 14 13:31:20.276373 containerd[1454]: time="2026-04-14T13:31:20.276367234Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Apr 14 13:31:20.276436 containerd[1454]: time="2026-04-14T13:31:20.276383265Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Apr 14 13:31:20.276436 containerd[1454]: time="2026-04-14T13:31:20.276402800Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 14 13:31:20.276436 containerd[1454]: time="2026-04-14T13:31:20.276419205Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 14 13:31:20.276517 containerd[1454]: time="2026-04-14T13:31:20.276436482Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 14 13:31:20.276517 containerd[1454]: time="2026-04-14T13:31:20.276453509Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 14 13:31:20.276517 containerd[1454]: time="2026-04-14T13:31:20.276471437Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 14 13:31:20.276517 containerd[1454]: time="2026-04-14T13:31:20.276493731Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 14 13:31:20.276517 containerd[1454]: time="2026-04-14T13:31:20.276511265Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 14 13:31:20.276663 containerd[1454]: time="2026-04-14T13:31:20.276525837Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 14 13:31:20.276663 containerd[1454]: time="2026-04-14T13:31:20.276551086Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276663 containerd[1454]: time="2026-04-14T13:31:20.276567904Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276663 containerd[1454]: time="2026-04-14T13:31:20.276582885Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276663 containerd[1454]: time="2026-04-14T13:31:20.276622583Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276663 containerd[1454]: time="2026-04-14T13:31:20.276639235Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276663 containerd[1454]: time="2026-04-14T13:31:20.276656113Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276670543Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276686929Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276703609Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276725840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276741181Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276757410Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276771585Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276794857Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Apr 14 13:31:20.276835 containerd[1454]: time="2026-04-14T13:31:20.276827956Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.276843943Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.276858458Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.276928020Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.276951787Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.276967828Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.276982713Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.276996670Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.277012168Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.277029088Z" level=info msg="NRI interface is disabled by configuration." Apr 14 13:31:20.277046 containerd[1454]: time="2026-04-14T13:31:20.277041812Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 14 13:31:20.277497 containerd[1454]: time="2026-04-14T13:31:20.277422782Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 14 13:31:20.277497 containerd[1454]: time="2026-04-14T13:31:20.277490570Z" level=info msg="Connect containerd service" Apr 14 13:31:20.277730 containerd[1454]: time="2026-04-14T13:31:20.277536588Z" level=info msg="using legacy CRI server" Apr 14 13:31:20.277730 containerd[1454]: time="2026-04-14T13:31:20.277545337Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 14 13:31:20.277730 containerd[1454]: time="2026-04-14T13:31:20.277706391Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 14 13:31:20.278391 containerd[1454]: time="2026-04-14T13:31:20.278355016Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278568249Z" level=info msg="Start subscribing containerd event" Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278619342Z" level=info msg="Start recovering state" Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278676973Z" level=info msg="Start event monitor" Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278717737Z" level=info msg="Start snapshots syncer" Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278732791Z" level=info msg="Start cni network conf syncer for default" Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278743260Z" level=info msg="Start streaming server" Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278829756Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278872481Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 14 13:31:20.278968 containerd[1454]: time="2026-04-14T13:31:20.278950413Z" level=info msg="containerd successfully booted in 0.037822s" Apr 14 13:31:20.279069 systemd[1]: Started containerd.service - containerd container runtime. Apr 14 13:31:20.547598 tar[1452]: linux-amd64/README.md Apr 14 13:31:20.559061 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 14 13:31:21.026826 systemd-networkd[1377]: eth0: Gained IPv6LL Apr 14 13:31:21.030167 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 14 13:31:21.032567 systemd[1]: Reached target network-online.target - Network is Online. Apr 14 13:31:21.050935 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Apr 14 13:31:21.054161 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 14 13:31:21.056855 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 14 13:31:21.074746 systemd[1]: coreos-metadata.service: Deactivated successfully. Apr 14 13:31:21.074907 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Apr 14 13:31:21.076992 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 14 13:31:21.079424 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 14 13:31:21.740286 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 14 13:31:21.742525 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 14 13:31:21.744665 systemd[1]: Startup finished in 841ms (kernel) + 5.375s (initrd) + 3.723s (userspace) = 9.940s. Apr 14 13:31:21.745117 (kubelet)[1538]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 14 13:31:22.152480 kubelet[1538]: E0414 13:31:22.152408 1538 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 14 13:31:22.154824 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 14 13:31:22.154948 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 14 13:31:26.377879 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 14 13:31:26.379070 systemd[1]: Started sshd@0-10.0.0.8:22-10.0.0.1:42744.service - OpenSSH per-connection server daemon (10.0.0.1:42744). Apr 14 13:31:26.423732 sshd[1552]: Accepted publickey for core from 10.0.0.1 port 42744 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:31:26.427825 sshd[1552]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:31:26.449323 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 14 13:31:26.457915 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 14 13:31:26.459591 systemd-logind[1442]: New session 1 of user core. Apr 14 13:31:26.468703 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 14 13:31:26.470564 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 14 13:31:26.478084 (systemd)[1556]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 14 13:31:26.570852 systemd[1556]: Queued start job for default target default.target. Apr 14 13:31:26.581010 systemd[1556]: Created slice app.slice - User Application Slice. Apr 14 13:31:26.581446 systemd[1556]: Reached target paths.target - Paths. Apr 14 13:31:26.581473 systemd[1556]: Reached target timers.target - Timers. Apr 14 13:31:26.582882 systemd[1556]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 14 13:31:26.596559 systemd[1556]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 14 13:31:26.596658 systemd[1556]: Reached target sockets.target - Sockets. Apr 14 13:31:26.596668 systemd[1556]: Reached target basic.target - Basic System. Apr 14 13:31:26.596694 systemd[1556]: Reached target default.target - Main User Target. Apr 14 13:31:26.596713 systemd[1556]: Startup finished in 112ms. Apr 14 13:31:26.596864 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 14 13:31:26.598086 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 14 13:31:26.661733 systemd[1]: Started sshd@1-10.0.0.8:22-10.0.0.1:42754.service - OpenSSH per-connection server daemon (10.0.0.1:42754). Apr 14 13:31:26.700590 sshd[1567]: Accepted publickey for core from 10.0.0.1 port 42754 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:31:26.702145 sshd[1567]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:31:26.708918 systemd-logind[1442]: New session 2 of user core. Apr 14 13:31:26.716507 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 14 13:31:26.775685 sshd[1567]: pam_unix(sshd:session): session closed for user core Apr 14 13:31:26.789749 systemd[1]: sshd@1-10.0.0.8:22-10.0.0.1:42754.service: Deactivated successfully. Apr 14 13:31:26.790964 systemd[1]: session-2.scope: Deactivated successfully. Apr 14 13:31:26.791989 systemd-logind[1442]: Session 2 logged out. Waiting for processes to exit. Apr 14 13:31:26.792971 systemd[1]: Started sshd@2-10.0.0.8:22-10.0.0.1:42770.service - OpenSSH per-connection server daemon (10.0.0.1:42770). Apr 14 13:31:26.793743 systemd-logind[1442]: Removed session 2. Apr 14 13:31:26.826318 sshd[1574]: Accepted publickey for core from 10.0.0.1 port 42770 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:31:26.827372 sshd[1574]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:31:26.831200 systemd-logind[1442]: New session 3 of user core. Apr 14 13:31:26.847026 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 14 13:31:26.899530 sshd[1574]: pam_unix(sshd:session): session closed for user core Apr 14 13:31:26.912815 systemd[1]: sshd@2-10.0.0.8:22-10.0.0.1:42770.service: Deactivated successfully. Apr 14 13:31:26.914034 systemd[1]: session-3.scope: Deactivated successfully. Apr 14 13:31:26.915074 systemd-logind[1442]: Session 3 logged out. Waiting for processes to exit. Apr 14 13:31:26.916045 systemd[1]: Started sshd@3-10.0.0.8:22-10.0.0.1:42778.service - OpenSSH per-connection server daemon (10.0.0.1:42778). Apr 14 13:31:26.916527 systemd-logind[1442]: Removed session 3. Apr 14 13:31:26.949879 sshd[1581]: Accepted publickey for core from 10.0.0.1 port 42778 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:31:26.950995 sshd[1581]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:31:26.954834 systemd-logind[1442]: New session 4 of user core. Apr 14 13:31:26.973818 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 14 13:31:27.032569 sshd[1581]: pam_unix(sshd:session): session closed for user core Apr 14 13:31:27.038909 systemd[1]: sshd@3-10.0.0.8:22-10.0.0.1:42778.service: Deactivated successfully. Apr 14 13:31:27.040530 systemd[1]: session-4.scope: Deactivated successfully. Apr 14 13:31:27.041488 systemd-logind[1442]: Session 4 logged out. Waiting for processes to exit. Apr 14 13:31:27.042425 systemd[1]: Started sshd@4-10.0.0.8:22-10.0.0.1:42794.service - OpenSSH per-connection server daemon (10.0.0.1:42794). Apr 14 13:31:27.043230 systemd-logind[1442]: Removed session 4. Apr 14 13:31:27.077356 sshd[1588]: Accepted publickey for core from 10.0.0.1 port 42794 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:31:27.078745 sshd[1588]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:31:27.084634 systemd-logind[1442]: New session 5 of user core. Apr 14 13:31:27.098656 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 14 13:31:27.163862 sudo[1591]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 14 13:31:27.164168 sudo[1591]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 14 13:31:27.182439 sudo[1591]: pam_unix(sudo:session): session closed for user root Apr 14 13:31:27.183964 sshd[1588]: pam_unix(sshd:session): session closed for user core Apr 14 13:31:27.190892 systemd[1]: sshd@4-10.0.0.8:22-10.0.0.1:42794.service: Deactivated successfully. Apr 14 13:31:27.192154 systemd[1]: session-5.scope: Deactivated successfully. Apr 14 13:31:27.193167 systemd-logind[1442]: Session 5 logged out. Waiting for processes to exit. Apr 14 13:31:27.194212 systemd[1]: Started sshd@5-10.0.0.8:22-10.0.0.1:42798.service - OpenSSH per-connection server daemon (10.0.0.1:42798). Apr 14 13:31:27.194969 systemd-logind[1442]: Removed session 5. Apr 14 13:31:27.227431 sshd[1596]: Accepted publickey for core from 10.0.0.1 port 42798 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:31:27.230464 sshd[1596]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:31:27.237587 systemd-logind[1442]: New session 6 of user core. Apr 14 13:31:27.251678 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 14 13:31:27.309599 sudo[1600]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 14 13:31:27.309828 sudo[1600]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 14 13:31:27.314141 sudo[1600]: pam_unix(sudo:session): session closed for user root Apr 14 13:31:27.320879 sudo[1599]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Apr 14 13:31:27.321101 sudo[1599]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 14 13:31:27.341800 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Apr 14 13:31:27.343439 auditctl[1603]: No rules Apr 14 13:31:27.343755 systemd[1]: audit-rules.service: Deactivated successfully. Apr 14 13:31:27.343974 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Apr 14 13:31:27.347098 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 14 13:31:27.376183 augenrules[1621]: No rules Apr 14 13:31:27.376959 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 14 13:31:27.377851 sudo[1599]: pam_unix(sudo:session): session closed for user root Apr 14 13:31:27.379963 sshd[1596]: pam_unix(sshd:session): session closed for user core Apr 14 13:31:27.392513 systemd[1]: sshd@5-10.0.0.8:22-10.0.0.1:42798.service: Deactivated successfully. Apr 14 13:31:27.393710 systemd[1]: session-6.scope: Deactivated successfully. Apr 14 13:31:27.394774 systemd-logind[1442]: Session 6 logged out. Waiting for processes to exit. Apr 14 13:31:27.396282 systemd[1]: Started sshd@6-10.0.0.8:22-10.0.0.1:42804.service - OpenSSH per-connection server daemon (10.0.0.1:42804). Apr 14 13:31:27.396841 systemd-logind[1442]: Removed session 6. Apr 14 13:31:27.430390 sshd[1629]: Accepted publickey for core from 10.0.0.1 port 42804 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:31:27.432691 sshd[1629]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:31:27.437628 systemd-logind[1442]: New session 7 of user core. Apr 14 13:31:27.452931 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 14 13:31:27.514000 sudo[1632]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 14 13:31:27.514486 sudo[1632]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 14 13:31:27.771572 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 14 13:31:27.771609 (dockerd)[1650]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 14 13:31:28.018919 dockerd[1650]: time="2026-04-14T13:31:28.018824519Z" level=info msg="Starting up" Apr 14 13:31:28.148459 dockerd[1650]: time="2026-04-14T13:31:28.148384484Z" level=info msg="Loading containers: start." Apr 14 13:31:28.358365 kernel: Initializing XFRM netlink socket Apr 14 13:31:28.437360 systemd-networkd[1377]: docker0: Link UP Apr 14 13:31:28.459051 dockerd[1650]: time="2026-04-14T13:31:28.458984892Z" level=info msg="Loading containers: done." Apr 14 13:31:28.471594 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1050351499-merged.mount: Deactivated successfully. Apr 14 13:31:28.472760 dockerd[1650]: time="2026-04-14T13:31:28.472688256Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 14 13:31:28.472826 dockerd[1650]: time="2026-04-14T13:31:28.472814108Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Apr 14 13:31:28.472918 dockerd[1650]: time="2026-04-14T13:31:28.472890349Z" level=info msg="Daemon has completed initialization" Apr 14 13:31:28.514593 dockerd[1650]: time="2026-04-14T13:31:28.514329870Z" level=info msg="API listen on /run/docker.sock" Apr 14 13:31:28.515479 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 14 13:31:28.933019 containerd[1454]: time="2026-04-14T13:31:28.932968137Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.35.3\"" Apr 14 13:31:29.696658 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount863161188.mount: Deactivated successfully. Apr 14 13:31:30.451557 containerd[1454]: time="2026-04-14T13:31:30.451403483Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.35.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:30.452002 containerd[1454]: time="2026-04-14T13:31:30.451951058Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.35.3: active requests=0, bytes read=27569134" Apr 14 13:31:30.453058 containerd[1454]: time="2026-04-14T13:31:30.453008215Z" level=info msg="ImageCreate event name:\"sha256:0f2b96c93465f04111c58c3fc41ad0ed2e16b5b3c4b6282b84dc951ad0ea4d66\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:30.455471 containerd[1454]: time="2026-04-14T13:31:30.455422901Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6c6e2571f98e738015a39ed21305ab4166a3e2873f9cc01d7fa58371cf0f5d30\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:30.456337 containerd[1454]: time="2026-04-14T13:31:30.456305274Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.35.3\" with image id \"sha256:0f2b96c93465f04111c58c3fc41ad0ed2e16b5b3c4b6282b84dc951ad0ea4d66\", repo tag \"registry.k8s.io/kube-apiserver:v1.35.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6c6e2571f98e738015a39ed21305ab4166a3e2873f9cc01d7fa58371cf0f5d30\", size \"27566295\" in 1.523284291s" Apr 14 13:31:30.456337 containerd[1454]: time="2026-04-14T13:31:30.456350504Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.35.3\" returns image reference \"sha256:0f2b96c93465f04111c58c3fc41ad0ed2e16b5b3c4b6282b84dc951ad0ea4d66\"" Apr 14 13:31:30.457009 containerd[1454]: time="2026-04-14T13:31:30.456983866Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.35.3\"" Apr 14 13:31:31.600678 containerd[1454]: time="2026-04-14T13:31:31.600598264Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.35.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:31.601683 containerd[1454]: time="2026-04-14T13:31:31.601618313Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.35.3: active requests=0, bytes read=21449527" Apr 14 13:31:31.602473 containerd[1454]: time="2026-04-14T13:31:31.602435832Z" level=info msg="ImageCreate event name:\"sha256:0eb506280f9bca2258673771e7029de0d5e92881f0fbaebd4a835e7e302b7d27\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:31.607908 containerd[1454]: time="2026-04-14T13:31:31.607772424Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:23a24aafa10831eb47477b0b31a525ee8a4a99d2c17251aac46c43be8201ec59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:31.608716 containerd[1454]: time="2026-04-14T13:31:31.608686226Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.35.3\" with image id \"sha256:0eb506280f9bca2258673771e7029de0d5e92881f0fbaebd4a835e7e302b7d27\", repo tag \"registry.k8s.io/kube-controller-manager:v1.35.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:23a24aafa10831eb47477b0b31a525ee8a4a99d2c17251aac46c43be8201ec59\", size \"23014443\" in 1.151671845s" Apr 14 13:31:31.608789 containerd[1454]: time="2026-04-14T13:31:31.608718739Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.35.3\" returns image reference \"sha256:0eb506280f9bca2258673771e7029de0d5e92881f0fbaebd4a835e7e302b7d27\"" Apr 14 13:31:31.609809 containerd[1454]: time="2026-04-14T13:31:31.609320968Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.35.3\"" Apr 14 13:31:32.338809 containerd[1454]: time="2026-04-14T13:31:32.338524774Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.35.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:32.339720 containerd[1454]: time="2026-04-14T13:31:32.339579185Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.35.3: active requests=0, bytes read=15548358" Apr 14 13:31:32.340635 containerd[1454]: time="2026-04-14T13:31:32.340591114Z" level=info msg="ImageCreate event name:\"sha256:87c9b0e4f80d3039b60fbfaf2a4d423e6a891df883a55adb58b8d5b37a4cb23c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:32.345457 containerd[1454]: time="2026-04-14T13:31:32.344186502Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:7070dff574916315268ab483f1088a107b1f3a8a1a87f3e3645933111ade7013\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:32.347033 containerd[1454]: time="2026-04-14T13:31:32.346914817Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.35.3\" with image id \"sha256:87c9b0e4f80d3039b60fbfaf2a4d423e6a891df883a55adb58b8d5b37a4cb23c\", repo tag \"registry.k8s.io/kube-scheduler:v1.35.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:7070dff574916315268ab483f1088a107b1f3a8a1a87f3e3645933111ade7013\", size \"17113292\" in 737.567932ms" Apr 14 13:31:32.347033 containerd[1454]: time="2026-04-14T13:31:32.346982991Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.35.3\" returns image reference \"sha256:87c9b0e4f80d3039b60fbfaf2a4d423e6a891df883a55adb58b8d5b37a4cb23c\"" Apr 14 13:31:32.347584 containerd[1454]: time="2026-04-14T13:31:32.347553411Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.35.3\"" Apr 14 13:31:32.398215 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 14 13:31:32.409743 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 14 13:31:32.589426 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 14 13:31:32.594971 (kubelet)[1873]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 14 13:31:32.692746 kubelet[1873]: E0414 13:31:32.692617 1873 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 14 13:31:32.696175 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 14 13:31:32.696398 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 14 13:31:32.881812 kernel: hrtimer: interrupt took 5297304 ns Apr 14 13:31:33.252550 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1579453558.mount: Deactivated successfully. Apr 14 13:31:33.465715 containerd[1454]: time="2026-04-14T13:31:33.465160296Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.35.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:33.466499 containerd[1454]: time="2026-04-14T13:31:33.466365376Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.35.3: active requests=0, bytes read=25685215" Apr 14 13:31:33.467775 containerd[1454]: time="2026-04-14T13:31:33.467681848Z" level=info msg="ImageCreate event name:\"sha256:53ed370019059b0cdce5a02a20f8aca81f977e34956368c7f1b7ce9709398b79\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:33.470434 containerd[1454]: time="2026-04-14T13:31:33.470338838Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8743aec6a360aedcb7a076cbecea367b072abe1bfade2e2098650df502e2bc89\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:33.470933 containerd[1454]: time="2026-04-14T13:31:33.470888268Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.35.3\" with image id \"sha256:53ed370019059b0cdce5a02a20f8aca81f977e34956368c7f1b7ce9709398b79\", repo tag \"registry.k8s.io/kube-proxy:v1.35.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:8743aec6a360aedcb7a076cbecea367b072abe1bfade2e2098650df502e2bc89\", size \"25684340\" in 1.123308349s" Apr 14 13:31:33.470933 containerd[1454]: time="2026-04-14T13:31:33.470917895Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.35.3\" returns image reference \"sha256:53ed370019059b0cdce5a02a20f8aca81f977e34956368c7f1b7ce9709398b79\"" Apr 14 13:31:33.471623 containerd[1454]: time="2026-04-14T13:31:33.471579300Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.13.1\"" Apr 14 13:31:33.918901 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3932166346.mount: Deactivated successfully. Apr 14 13:31:34.608873 containerd[1454]: time="2026-04-14T13:31:34.608801306Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.13.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:34.609577 containerd[1454]: time="2026-04-14T13:31:34.609438534Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.13.1: active requests=0, bytes read=23555980" Apr 14 13:31:34.610568 containerd[1454]: time="2026-04-14T13:31:34.610520846Z" level=info msg="ImageCreate event name:\"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:34.612773 containerd[1454]: time="2026-04-14T13:31:34.612731249Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9b9128672209474da07c91439bf15ed704ae05ad918dd6454e5b6ae14e35fee6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:34.613866 containerd[1454]: time="2026-04-14T13:31:34.613833811Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.13.1\" with image id \"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\", repo tag \"registry.k8s.io/coredns/coredns:v1.13.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9b9128672209474da07c91439bf15ed704ae05ad918dd6454e5b6ae14e35fee6\", size \"23553139\" in 1.142223262s" Apr 14 13:31:34.613914 containerd[1454]: time="2026-04-14T13:31:34.613872600Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.13.1\" returns image reference \"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\"" Apr 14 13:31:34.614744 containerd[1454]: time="2026-04-14T13:31:34.614713210Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Apr 14 13:31:35.022011 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3281864048.mount: Deactivated successfully. Apr 14 13:31:35.030990 containerd[1454]: time="2026-04-14T13:31:35.030906507Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:35.031519 containerd[1454]: time="2026-04-14T13:31:35.031464329Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=321150" Apr 14 13:31:35.032512 containerd[1454]: time="2026-04-14T13:31:35.032475877Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:35.035085 containerd[1454]: time="2026-04-14T13:31:35.035001089Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:35.035644 containerd[1454]: time="2026-04-14T13:31:35.035598262Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 420.851092ms" Apr 14 13:31:35.035644 containerd[1454]: time="2026-04-14T13:31:35.035630243Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Apr 14 13:31:35.036188 containerd[1454]: time="2026-04-14T13:31:35.036143418Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.6-0\"" Apr 14 13:31:35.439513 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount186980173.mount: Deactivated successfully. Apr 14 13:31:36.167425 containerd[1454]: time="2026-04-14T13:31:36.167130318Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.6-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:36.168446 containerd[1454]: time="2026-04-14T13:31:36.168403721Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.6-0: active requests=0, bytes read=23643431" Apr 14 13:31:36.169501 containerd[1454]: time="2026-04-14T13:31:36.169428334Z" level=info msg="ImageCreate event name:\"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:36.175323 containerd[1454]: time="2026-04-14T13:31:36.175179290Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:60a30b5d81b2217555e2cfb9537f655b7ba97220b99c39ee2e162a7127225890\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:36.176110 containerd[1454]: time="2026-04-14T13:31:36.176074948Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.6-0\" with image id \"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\", repo tag \"registry.k8s.io/etcd:3.6.6-0\", repo digest \"registry.k8s.io/etcd@sha256:60a30b5d81b2217555e2cfb9537f655b7ba97220b99c39ee2e162a7127225890\", size \"23641797\" in 1.139910804s" Apr 14 13:31:36.176110 containerd[1454]: time="2026-04-14T13:31:36.176097835Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.6-0\" returns image reference \"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\"" Apr 14 13:31:37.426939 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 14 13:31:37.442519 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 14 13:31:37.470772 systemd[1]: Reloading requested from client PID 2043 ('systemctl') (unit session-7.scope)... Apr 14 13:31:37.470795 systemd[1]: Reloading... Apr 14 13:31:37.542298 zram_generator::config[2082]: No configuration found. Apr 14 13:31:37.645631 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 14 13:31:37.700069 systemd[1]: Reloading finished in 229 ms. Apr 14 13:31:37.745677 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Apr 14 13:31:37.745734 systemd[1]: kubelet.service: Failed with result 'signal'. Apr 14 13:31:37.745915 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 14 13:31:37.748438 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 14 13:31:37.879707 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 14 13:31:37.883757 (kubelet)[2131]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 14 13:31:37.966107 kubelet[2131]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 14 13:31:38.044478 kubelet[2131]: I0414 13:31:38.044389 2131 server.go:525] "Kubelet version" kubeletVersion="v1.35.1" Apr 14 13:31:38.044478 kubelet[2131]: I0414 13:31:38.044433 2131 server.go:527] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 14 13:31:38.044478 kubelet[2131]: I0414 13:31:38.044461 2131 watchdog_linux.go:95] "Systemd watchdog is not enabled" Apr 14 13:31:38.044478 kubelet[2131]: I0414 13:31:38.044465 2131 watchdog_linux.go:138] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 14 13:31:38.044760 kubelet[2131]: I0414 13:31:38.044712 2131 server.go:951] "Client rotation is on, will bootstrap in background" Apr 14 13:31:38.073997 kubelet[2131]: E0414 13:31:38.073848 2131 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.8:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.8:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 14 13:31:38.077234 kubelet[2131]: I0414 13:31:38.077160 2131 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 14 13:31:38.079930 kubelet[2131]: E0414 13:31:38.079892 2131 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 14 13:31:38.079930 kubelet[2131]: I0414 13:31:38.079933 2131 server.go:1395] "CRI implementation should be updated to support RuntimeConfig. Falling back to using cgroupDriver from kubelet config." Apr 14 13:31:38.083642 kubelet[2131]: I0414 13:31:38.083598 2131 server.go:775] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Apr 14 13:31:38.084590 kubelet[2131]: I0414 13:31:38.084538 2131 container_manager_linux.go:272] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 14 13:31:38.084770 kubelet[2131]: I0414 13:31:38.084570 2131 container_manager_linux.go:277] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 14 13:31:38.084770 kubelet[2131]: I0414 13:31:38.084753 2131 topology_manager.go:143] "Creating topology manager with none policy" Apr 14 13:31:38.084770 kubelet[2131]: I0414 13:31:38.084763 2131 container_manager_linux.go:308] "Creating device plugin manager" Apr 14 13:31:38.084921 kubelet[2131]: I0414 13:31:38.084833 2131 container_manager_linux.go:317] "Creating Dynamic Resource Allocation (DRA) manager" Apr 14 13:31:38.086547 kubelet[2131]: I0414 13:31:38.086513 2131 state_mem.go:41] "Initialized" logger="CPUManager state memory" Apr 14 13:31:38.086723 kubelet[2131]: I0414 13:31:38.086696 2131 kubelet.go:482] "Attempting to sync node with API server" Apr 14 13:31:38.086723 kubelet[2131]: I0414 13:31:38.086716 2131 kubelet.go:383] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 14 13:31:38.086765 kubelet[2131]: I0414 13:31:38.086735 2131 kubelet.go:394] "Adding apiserver pod source" Apr 14 13:31:38.086765 kubelet[2131]: I0414 13:31:38.086745 2131 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 14 13:31:38.088786 kubelet[2131]: I0414 13:31:38.088735 2131 kuberuntime_manager.go:294] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 14 13:31:38.091769 kubelet[2131]: I0414 13:31:38.091738 2131 kubelet.go:943] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 14 13:31:38.091806 kubelet[2131]: I0414 13:31:38.091793 2131 kubelet.go:970] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Apr 14 13:31:38.091914 kubelet[2131]: W0414 13:31:38.091861 2131 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 14 13:31:38.096285 kubelet[2131]: I0414 13:31:38.096083 2131 server.go:1257] "Started kubelet" Apr 14 13:31:38.096285 kubelet[2131]: I0414 13:31:38.096209 2131 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 14 13:31:38.096372 kubelet[2131]: I0414 13:31:38.096332 2131 server_v1.go:49] "podresources" method="list" useActivePods=true Apr 14 13:31:38.096610 kubelet[2131]: I0414 13:31:38.096210 2131 server.go:182] "Starting to listen" address="0.0.0.0" port=10250 Apr 14 13:31:38.096610 kubelet[2131]: I0414 13:31:38.096564 2131 server.go:254] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 14 13:31:38.097683 kubelet[2131]: I0414 13:31:38.097643 2131 fs_resource_analyzer.go:69] "Starting FS ResourceAnalyzer" Apr 14 13:31:38.098568 kubelet[2131]: I0414 13:31:38.098132 2131 server.go:317] "Adding debug handlers to kubelet server" Apr 14 13:31:38.100506 kubelet[2131]: I0414 13:31:38.100465 2131 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 14 13:31:38.101610 kubelet[2131]: I0414 13:31:38.101311 2131 volume_manager.go:311] "Starting Kubelet Volume Manager" Apr 14 13:31:38.102975 kubelet[2131]: I0414 13:31:38.102914 2131 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 14 13:31:38.102975 kubelet[2131]: I0414 13:31:38.102965 2131 reconciler.go:29] "Reconciler: start to sync state" Apr 14 13:31:38.104668 kubelet[2131]: E0414 13:31:38.104436 2131 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 14 13:31:38.105037 kubelet[2131]: E0414 13:31:38.104986 2131 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.8:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.8:6443: connect: connection refused" interval="200ms" Apr 14 13:31:38.106015 kubelet[2131]: E0414 13:31:38.105045 2131 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.8:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.8:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18a63c62cd7a4408 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-14 13:31:38.096043016 +0000 UTC m=+0.201407961,LastTimestamp:2026-04-14 13:31:38.096043016 +0000 UTC m=+0.201407961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 14 13:31:38.106166 kubelet[2131]: I0414 13:31:38.106148 2131 factory.go:223] Registration of the containerd container factory successfully Apr 14 13:31:38.106166 kubelet[2131]: I0414 13:31:38.106165 2131 factory.go:223] Registration of the systemd container factory successfully Apr 14 13:31:38.106333 kubelet[2131]: I0414 13:31:38.106301 2131 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 14 13:31:38.109245 kubelet[2131]: E0414 13:31:38.107365 2131 kubelet.go:1656] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 14 13:31:38.128784 kubelet[2131]: I0414 13:31:38.128761 2131 cpu_manager.go:225] "Starting" policy="none" Apr 14 13:31:38.128784 kubelet[2131]: I0414 13:31:38.128777 2131 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Apr 14 13:31:38.128784 kubelet[2131]: I0414 13:31:38.128791 2131 state_mem.go:41] "Initialized" logger="CPUManager state checkpoint.CPUManager state memory" Apr 14 13:31:38.130871 kubelet[2131]: I0414 13:31:38.130847 2131 policy_none.go:50] "Start" Apr 14 13:31:38.130871 kubelet[2131]: I0414 13:31:38.130869 2131 memory_manager.go:187] "Starting memorymanager" policy="None" Apr 14 13:31:38.130967 kubelet[2131]: I0414 13:31:38.130878 2131 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Apr 14 13:31:38.132530 kubelet[2131]: I0414 13:31:38.132514 2131 policy_none.go:44] "Start" Apr 14 13:31:38.136093 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 14 13:31:38.138176 kubelet[2131]: I0414 13:31:38.138131 2131 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Apr 14 13:31:38.139634 kubelet[2131]: I0414 13:31:38.139590 2131 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Apr 14 13:31:38.139634 kubelet[2131]: I0414 13:31:38.139617 2131 status_manager.go:249] "Starting to sync pod status with apiserver" Apr 14 13:31:38.139724 kubelet[2131]: I0414 13:31:38.139642 2131 kubelet.go:2501] "Starting kubelet main sync loop" Apr 14 13:31:38.139724 kubelet[2131]: E0414 13:31:38.139675 2131 kubelet.go:2525] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 14 13:31:38.144410 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 14 13:31:38.147335 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 14 13:31:38.161275 kubelet[2131]: E0414 13:31:38.161209 2131 manager.go:525] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 14 13:31:38.161487 kubelet[2131]: I0414 13:31:38.161416 2131 eviction_manager.go:194] "Eviction manager: starting control loop" Apr 14 13:31:38.161487 kubelet[2131]: I0414 13:31:38.161429 2131 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 14 13:31:38.161653 kubelet[2131]: I0414 13:31:38.161622 2131 plugin_manager.go:121] "Starting Kubelet Plugin Manager" Apr 14 13:31:38.162359 kubelet[2131]: E0414 13:31:38.162272 2131 eviction_manager.go:272] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 14 13:31:38.162359 kubelet[2131]: E0414 13:31:38.162299 2131 eviction_manager.go:297] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 14 13:31:38.252319 systemd[1]: Created slice kubepods-burstable-pod3566c1d7ed03bb3c60facf009a5678dd.slice - libcontainer container kubepods-burstable-pod3566c1d7ed03bb3c60facf009a5678dd.slice. Apr 14 13:31:38.262604 kubelet[2131]: E0414 13:31:38.262452 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:38.264560 kubelet[2131]: I0414 13:31:38.264523 2131 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Apr 14 13:31:38.264880 kubelet[2131]: E0414 13:31:38.264845 2131 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.8:6443/api/v1/nodes\": dial tcp 10.0.0.8:6443: connect: connection refused" node="localhost" Apr 14 13:31:38.265447 systemd[1]: Created slice kubepods-burstable-podeb404586bd2117467da4e586f8d427df.slice - libcontainer container kubepods-burstable-podeb404586bd2117467da4e586f8d427df.slice. Apr 14 13:31:38.276040 kubelet[2131]: E0414 13:31:38.275920 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:38.278738 systemd[1]: Created slice kubepods-burstable-podbd70d524e6bc561f2082b467706799ed.slice - libcontainer container kubepods-burstable-podbd70d524e6bc561f2082b467706799ed.slice. Apr 14 13:31:38.280559 kubelet[2131]: E0414 13:31:38.280504 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:38.305961 kubelet[2131]: E0414 13:31:38.305782 2131 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.8:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.8:6443: connect: connection refused" interval="400ms" Apr 14 13:31:38.405266 kubelet[2131]: I0414 13:31:38.405137 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3566c1d7ed03bb3c60facf009a5678dd-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"3566c1d7ed03bb3c60facf009a5678dd\") " pod="kube-system/kube-scheduler-localhost" Apr 14 13:31:38.405266 kubelet[2131]: I0414 13:31:38.405249 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/eb404586bd2117467da4e586f8d427df-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"eb404586bd2117467da4e586f8d427df\") " pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:38.405266 kubelet[2131]: I0414 13:31:38.405275 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/eb404586bd2117467da4e586f8d427df-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"eb404586bd2117467da4e586f8d427df\") " pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:38.405266 kubelet[2131]: I0414 13:31:38.405288 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/eb404586bd2117467da4e586f8d427df-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"eb404586bd2117467da4e586f8d427df\") " pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:38.406163 kubelet[2131]: I0414 13:31:38.405305 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:38.406163 kubelet[2131]: I0414 13:31:38.405316 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:38.406163 kubelet[2131]: I0414 13:31:38.405327 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:38.406163 kubelet[2131]: I0414 13:31:38.405338 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:38.406163 kubelet[2131]: I0414 13:31:38.405350 2131 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:38.468499 kubelet[2131]: I0414 13:31:38.468413 2131 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Apr 14 13:31:38.468742 kubelet[2131]: E0414 13:31:38.468699 2131 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.8:6443/api/v1/nodes\": dial tcp 10.0.0.8:6443: connect: connection refused" node="localhost" Apr 14 13:31:38.573631 kubelet[2131]: E0414 13:31:38.571978 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:38.575878 containerd[1454]: time="2026-04-14T13:31:38.575760445Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:3566c1d7ed03bb3c60facf009a5678dd,Namespace:kube-system,Attempt:0,}" Apr 14 13:31:38.590067 kubelet[2131]: E0414 13:31:38.589972 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:38.599818 kubelet[2131]: E0414 13:31:38.598780 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:38.604358 containerd[1454]: time="2026-04-14T13:31:38.604291010Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:eb404586bd2117467da4e586f8d427df,Namespace:kube-system,Attempt:0,}" Apr 14 13:31:38.604800 containerd[1454]: time="2026-04-14T13:31:38.604740388Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:bd70d524e6bc561f2082b467706799ed,Namespace:kube-system,Attempt:0,}" Apr 14 13:31:38.712339 kubelet[2131]: E0414 13:31:38.712281 2131 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.8:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.8:6443: connect: connection refused" interval="800ms" Apr 14 13:31:38.871676 kubelet[2131]: I0414 13:31:38.871533 2131 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Apr 14 13:31:38.872060 kubelet[2131]: E0414 13:31:38.871985 2131 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.8:6443/api/v1/nodes\": dial tcp 10.0.0.8:6443: connect: connection refused" node="localhost" Apr 14 13:31:39.075988 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount444709749.mount: Deactivated successfully. Apr 14 13:31:39.080533 containerd[1454]: time="2026-04-14T13:31:39.080492684Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 14 13:31:39.081139 containerd[1454]: time="2026-04-14T13:31:39.081099951Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=311988" Apr 14 13:31:39.084037 containerd[1454]: time="2026-04-14T13:31:39.083996440Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 14 13:31:39.087133 containerd[1454]: time="2026-04-14T13:31:39.087061820Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 14 13:31:39.088589 containerd[1454]: time="2026-04-14T13:31:39.088521330Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 14 13:31:39.089380 containerd[1454]: time="2026-04-14T13:31:39.089301452Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 14 13:31:39.089953 containerd[1454]: time="2026-04-14T13:31:39.089921611Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 14 13:31:39.091279 containerd[1454]: time="2026-04-14T13:31:39.091253368Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 14 13:31:39.092315 containerd[1454]: time="2026-04-14T13:31:39.092279662Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 487.910954ms" Apr 14 13:31:39.093165 containerd[1454]: time="2026-04-14T13:31:39.093119676Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 517.253356ms" Apr 14 13:31:39.095698 containerd[1454]: time="2026-04-14T13:31:39.095636449Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 490.839522ms" Apr 14 13:31:39.289687 containerd[1454]: time="2026-04-14T13:31:39.289413861Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:31:39.289687 containerd[1454]: time="2026-04-14T13:31:39.289495566Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:31:39.289687 containerd[1454]: time="2026-04-14T13:31:39.289503718Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:39.289687 containerd[1454]: time="2026-04-14T13:31:39.289613823Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:39.289687 containerd[1454]: time="2026-04-14T13:31:39.289625713Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:31:39.289687 containerd[1454]: time="2026-04-14T13:31:39.289659799Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:31:39.289687 containerd[1454]: time="2026-04-14T13:31:39.289672317Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:39.289954 containerd[1454]: time="2026-04-14T13:31:39.289743243Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:39.540665 kubelet[2131]: E0414 13:31:39.518930 2131 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.8:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.8:6443: connect: connection refused" interval="1.6s" Apr 14 13:31:39.645624 systemd[1]: Started cri-containerd-4c352d324ecfc05bed938f1e0a582471df0faba5c940da7b81a29302b57f35d3.scope - libcontainer container 4c352d324ecfc05bed938f1e0a582471df0faba5c940da7b81a29302b57f35d3. Apr 14 13:31:39.652690 systemd[1]: Started cri-containerd-2116da99753becd67103f25f39e416f07238c91e0b0f7970ee0e643a118777de.scope - libcontainer container 2116da99753becd67103f25f39e416f07238c91e0b0f7970ee0e643a118777de. Apr 14 13:31:39.665552 containerd[1454]: time="2026-04-14T13:31:39.665301782Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:31:39.665552 containerd[1454]: time="2026-04-14T13:31:39.665526213Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:31:39.665552 containerd[1454]: time="2026-04-14T13:31:39.665559011Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:39.665826 containerd[1454]: time="2026-04-14T13:31:39.665767937Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:39.683612 kubelet[2131]: I0414 13:31:39.683470 2131 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Apr 14 13:31:39.687427 kubelet[2131]: E0414 13:31:39.687316 2131 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.0.0.8:6443/api/v1/nodes\": dial tcp 10.0.0.8:6443: connect: connection refused" node="localhost" Apr 14 13:31:39.701824 systemd[1]: Started cri-containerd-125e1a7f35087de6dccedae06221ec56237ecf1bd6430003b4efb951005f1589.scope - libcontainer container 125e1a7f35087de6dccedae06221ec56237ecf1bd6430003b4efb951005f1589. Apr 14 13:31:39.725717 containerd[1454]: time="2026-04-14T13:31:39.725639302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:eb404586bd2117467da4e586f8d427df,Namespace:kube-system,Attempt:0,} returns sandbox id \"2116da99753becd67103f25f39e416f07238c91e0b0f7970ee0e643a118777de\"" Apr 14 13:31:39.727682 containerd[1454]: time="2026-04-14T13:31:39.727649006Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:3566c1d7ed03bb3c60facf009a5678dd,Namespace:kube-system,Attempt:0,} returns sandbox id \"4c352d324ecfc05bed938f1e0a582471df0faba5c940da7b81a29302b57f35d3\"" Apr 14 13:31:39.730060 kubelet[2131]: E0414 13:31:39.730024 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:39.730684 kubelet[2131]: E0414 13:31:39.730654 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:39.741511 containerd[1454]: time="2026-04-14T13:31:39.741454990Z" level=info msg="CreateContainer within sandbox \"4c352d324ecfc05bed938f1e0a582471df0faba5c940da7b81a29302b57f35d3\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 14 13:31:39.743165 containerd[1454]: time="2026-04-14T13:31:39.743115457Z" level=info msg="CreateContainer within sandbox \"2116da99753becd67103f25f39e416f07238c91e0b0f7970ee0e643a118777de\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 14 13:31:39.755903 containerd[1454]: time="2026-04-14T13:31:39.755861955Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:bd70d524e6bc561f2082b467706799ed,Namespace:kube-system,Attempt:0,} returns sandbox id \"125e1a7f35087de6dccedae06221ec56237ecf1bd6430003b4efb951005f1589\"" Apr 14 13:31:39.757440 kubelet[2131]: E0414 13:31:39.756978 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:39.764595 containerd[1454]: time="2026-04-14T13:31:39.764448487Z" level=info msg="CreateContainer within sandbox \"125e1a7f35087de6dccedae06221ec56237ecf1bd6430003b4efb951005f1589\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 14 13:31:39.764713 containerd[1454]: time="2026-04-14T13:31:39.764683322Z" level=info msg="CreateContainer within sandbox \"2116da99753becd67103f25f39e416f07238c91e0b0f7970ee0e643a118777de\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0a89ff6aadcefad9103e98a6675227c982d997d29561ca2641e288c7ea694403\"" Apr 14 13:31:39.765815 containerd[1454]: time="2026-04-14T13:31:39.765745815Z" level=info msg="StartContainer for \"0a89ff6aadcefad9103e98a6675227c982d997d29561ca2641e288c7ea694403\"" Apr 14 13:31:39.770363 containerd[1454]: time="2026-04-14T13:31:39.770320132Z" level=info msg="CreateContainer within sandbox \"4c352d324ecfc05bed938f1e0a582471df0faba5c940da7b81a29302b57f35d3\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"ace5cc85eea7199c216cf0666c60682a7da2c7af8712220451bf0d6137bcd5fe\"" Apr 14 13:31:39.772165 containerd[1454]: time="2026-04-14T13:31:39.772065268Z" level=info msg="StartContainer for \"ace5cc85eea7199c216cf0666c60682a7da2c7af8712220451bf0d6137bcd5fe\"" Apr 14 13:31:39.793933 containerd[1454]: time="2026-04-14T13:31:39.793786819Z" level=info msg="CreateContainer within sandbox \"125e1a7f35087de6dccedae06221ec56237ecf1bd6430003b4efb951005f1589\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"12de87055efb3a144f23c1f4a8a8f7e03be30477bdaa313cf357bfd011e87b83\"" Apr 14 13:31:39.794949 containerd[1454]: time="2026-04-14T13:31:39.794889273Z" level=info msg="StartContainer for \"12de87055efb3a144f23c1f4a8a8f7e03be30477bdaa313cf357bfd011e87b83\"" Apr 14 13:31:40.131700 systemd[1]: Started cri-containerd-ace5cc85eea7199c216cf0666c60682a7da2c7af8712220451bf0d6137bcd5fe.scope - libcontainer container ace5cc85eea7199c216cf0666c60682a7da2c7af8712220451bf0d6137bcd5fe. Apr 14 13:31:40.143994 kubelet[2131]: E0414 13:31:40.143718 2131 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.8:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.8:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 14 13:31:40.155809 systemd[1]: Started cri-containerd-12de87055efb3a144f23c1f4a8a8f7e03be30477bdaa313cf357bfd011e87b83.scope - libcontainer container 12de87055efb3a144f23c1f4a8a8f7e03be30477bdaa313cf357bfd011e87b83. Apr 14 13:31:40.160154 systemd[1]: Started cri-containerd-0a89ff6aadcefad9103e98a6675227c982d997d29561ca2641e288c7ea694403.scope - libcontainer container 0a89ff6aadcefad9103e98a6675227c982d997d29561ca2641e288c7ea694403. Apr 14 13:31:40.216122 containerd[1454]: time="2026-04-14T13:31:40.216054077Z" level=info msg="StartContainer for \"ace5cc85eea7199c216cf0666c60682a7da2c7af8712220451bf0d6137bcd5fe\" returns successfully" Apr 14 13:31:40.243108 containerd[1454]: time="2026-04-14T13:31:40.243047496Z" level=info msg="StartContainer for \"0a89ff6aadcefad9103e98a6675227c982d997d29561ca2641e288c7ea694403\" returns successfully" Apr 14 13:31:40.243327 containerd[1454]: time="2026-04-14T13:31:40.243075962Z" level=info msg="StartContainer for \"12de87055efb3a144f23c1f4a8a8f7e03be30477bdaa313cf357bfd011e87b83\" returns successfully" Apr 14 13:31:41.212619 kubelet[2131]: E0414 13:31:41.212564 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:41.213563 kubelet[2131]: E0414 13:31:41.212735 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:41.226928 kubelet[2131]: E0414 13:31:41.226886 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:41.227057 kubelet[2131]: E0414 13:31:41.227048 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:41.268479 kubelet[2131]: E0414 13:31:41.268399 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:41.268800 kubelet[2131]: E0414 13:31:41.268729 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:41.331257 kubelet[2131]: I0414 13:31:41.331199 2131 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Apr 14 13:31:42.282057 kubelet[2131]: E0414 13:31:42.281998 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:42.282594 kubelet[2131]: E0414 13:31:42.282148 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:42.282594 kubelet[2131]: E0414 13:31:42.282423 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:42.282594 kubelet[2131]: E0414 13:31:42.282507 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:42.282720 kubelet[2131]: E0414 13:31:42.282682 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:42.282793 kubelet[2131]: E0414 13:31:42.282774 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:43.291987 kubelet[2131]: E0414 13:31:43.291912 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:43.292717 kubelet[2131]: E0414 13:31:43.292215 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:43.294316 kubelet[2131]: E0414 13:31:43.293462 2131 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 14 13:31:43.294316 kubelet[2131]: E0414 13:31:43.293575 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:43.986493 kubelet[2131]: E0414 13:31:43.986443 2131 nodelease.go:50] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Apr 14 13:31:44.078302 kubelet[2131]: I0414 13:31:44.073570 2131 kubelet_node_status.go:77] "Successfully registered node" node="localhost" Apr 14 13:31:44.096901 kubelet[2131]: I0414 13:31:44.096774 2131 apiserver.go:52] "Watching apiserver" Apr 14 13:31:44.123555 kubelet[2131]: I0414 13:31:44.123441 2131 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:44.276125 kubelet[2131]: E0414 13:31:44.269353 2131 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.18a63c62cd7a4408 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-14 13:31:38.096043016 +0000 UTC m=+0.201407961,LastTimestamp:2026-04-14 13:31:38.096043016 +0000 UTC m=+0.201407961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 14 13:31:44.393902 kubelet[2131]: I0414 13:31:44.391570 2131 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:44.404308 kubelet[2131]: I0414 13:31:44.404247 2131 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 14 13:31:44.420626 kubelet[2131]: E0414 13:31:44.420536 2131 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:44.421492 kubelet[2131]: E0414 13:31:44.421172 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:44.421492 kubelet[2131]: E0414 13:31:44.421380 2131 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:44.421492 kubelet[2131]: I0414 13:31:44.421397 2131 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:44.431436 kubelet[2131]: E0414 13:31:44.430874 2131 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:44.431436 kubelet[2131]: I0414 13:31:44.431027 2131 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Apr 14 13:31:44.439948 kubelet[2131]: E0414 13:31:44.439600 2131 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Apr 14 13:31:47.948635 kubelet[2131]: I0414 13:31:47.948285 2131 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:47.966755 kubelet[2131]: E0414 13:31:47.966681 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:48.424264 kubelet[2131]: E0414 13:31:48.424191 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:48.875312 systemd[1]: Reloading requested from client PID 2422 ('systemctl') (unit session-7.scope)... Apr 14 13:31:48.875342 systemd[1]: Reloading... Apr 14 13:31:48.962777 kubelet[2131]: I0414 13:31:48.961944 2131 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Apr 14 13:31:48.979615 kubelet[2131]: I0414 13:31:48.979384 2131 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.979369267 podStartE2EDuration="1.979369267s" podCreationTimestamp="2026-04-14 13:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-14 13:31:48.383422601 +0000 UTC m=+10.488787554" watchObservedRunningTime="2026-04-14 13:31:48.979369267 +0000 UTC m=+11.084734211" Apr 14 13:31:48.980509 kubelet[2131]: E0414 13:31:48.980126 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:49.350303 zram_generator::config[2461]: No configuration found. Apr 14 13:31:49.428133 kubelet[2131]: E0414 13:31:49.427996 2131 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:49.490052 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 14 13:31:49.590762 systemd[1]: Reloading finished in 715 ms. Apr 14 13:31:49.736909 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 14 13:31:49.761548 systemd[1]: kubelet.service: Deactivated successfully. Apr 14 13:31:49.761772 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 14 13:31:49.762286 systemd[1]: kubelet.service: Consumed 2.273s CPU time, 129.6M memory peak, 0B memory swap peak. Apr 14 13:31:49.779687 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 14 13:31:50.040144 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 14 13:31:50.047013 (kubelet)[2506]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 14 13:31:50.312505 kubelet[2506]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 14 13:31:50.320417 kubelet[2506]: I0414 13:31:50.320309 2506 server.go:525] "Kubelet version" kubeletVersion="v1.35.1" Apr 14 13:31:50.320417 kubelet[2506]: I0414 13:31:50.320391 2506 server.go:527] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 14 13:31:50.320417 kubelet[2506]: I0414 13:31:50.320398 2506 watchdog_linux.go:95] "Systemd watchdog is not enabled" Apr 14 13:31:50.320417 kubelet[2506]: I0414 13:31:50.320404 2506 watchdog_linux.go:138] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 14 13:31:50.320637 kubelet[2506]: I0414 13:31:50.320590 2506 server.go:951] "Client rotation is on, will bootstrap in background" Apr 14 13:31:50.321638 kubelet[2506]: I0414 13:31:50.321600 2506 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Apr 14 13:31:50.324271 kubelet[2506]: I0414 13:31:50.323704 2506 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 14 13:31:50.330264 kubelet[2506]: E0414 13:31:50.329584 2506 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 14 13:31:50.330264 kubelet[2506]: I0414 13:31:50.329692 2506 server.go:1395] "CRI implementation should be updated to support RuntimeConfig. Falling back to using cgroupDriver from kubelet config." Apr 14 13:31:50.334207 kubelet[2506]: I0414 13:31:50.334090 2506 server.go:775] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Apr 14 13:31:50.334432 kubelet[2506]: I0414 13:31:50.334384 2506 container_manager_linux.go:272] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 14 13:31:50.335524 kubelet[2506]: I0414 13:31:50.334428 2506 container_manager_linux.go:277] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 14 13:31:50.335524 kubelet[2506]: I0414 13:31:50.335071 2506 topology_manager.go:143] "Creating topology manager with none policy" Apr 14 13:31:50.335524 kubelet[2506]: I0414 13:31:50.335086 2506 container_manager_linux.go:308] "Creating device plugin manager" Apr 14 13:31:50.335524 kubelet[2506]: I0414 13:31:50.335143 2506 container_manager_linux.go:317] "Creating Dynamic Resource Allocation (DRA) manager" Apr 14 13:31:50.335960 kubelet[2506]: I0414 13:31:50.335917 2506 state_mem.go:41] "Initialized" logger="CPUManager state memory" Apr 14 13:31:50.336128 kubelet[2506]: I0414 13:31:50.336089 2506 kubelet.go:482] "Attempting to sync node with API server" Apr 14 13:31:50.339247 kubelet[2506]: I0414 13:31:50.336114 2506 kubelet.go:383] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 14 13:31:50.339247 kubelet[2506]: I0414 13:31:50.336485 2506 kubelet.go:394] "Adding apiserver pod source" Apr 14 13:31:50.339247 kubelet[2506]: I0414 13:31:50.336493 2506 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 14 13:31:50.339866 kubelet[2506]: I0414 13:31:50.339808 2506 kuberuntime_manager.go:294] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 14 13:31:50.340651 kubelet[2506]: I0414 13:31:50.340583 2506 kubelet.go:943] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 14 13:31:50.340651 kubelet[2506]: I0414 13:31:50.340627 2506 kubelet.go:970] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Apr 14 13:31:50.348175 kubelet[2506]: I0414 13:31:50.348160 2506 server.go:1257] "Started kubelet" Apr 14 13:31:50.348824 kubelet[2506]: I0414 13:31:50.348570 2506 server.go:182] "Starting to listen" address="0.0.0.0" port=10250 Apr 14 13:31:50.352788 kubelet[2506]: I0414 13:31:50.352746 2506 fs_resource_analyzer.go:69] "Starting FS ResourceAnalyzer" Apr 14 13:31:50.353107 kubelet[2506]: I0414 13:31:50.353068 2506 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 14 13:31:50.353168 kubelet[2506]: I0414 13:31:50.353161 2506 server_v1.go:49] "podresources" method="list" useActivePods=true Apr 14 13:31:50.353462 kubelet[2506]: I0414 13:31:50.353450 2506 server.go:254] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 14 13:31:50.366526 kubelet[2506]: I0414 13:31:50.366440 2506 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 14 13:31:50.367825 kubelet[2506]: I0414 13:31:50.367811 2506 server.go:317] "Adding debug handlers to kubelet server" Apr 14 13:31:50.373554 kubelet[2506]: I0414 13:31:50.373499 2506 volume_manager.go:311] "Starting Kubelet Volume Manager" Apr 14 13:31:50.373816 kubelet[2506]: E0414 13:31:50.373695 2506 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 14 13:31:50.374185 kubelet[2506]: I0414 13:31:50.374149 2506 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 14 13:31:50.374367 kubelet[2506]: I0414 13:31:50.374337 2506 reconciler.go:29] "Reconciler: start to sync state" Apr 14 13:31:50.380482 kubelet[2506]: I0414 13:31:50.380465 2506 factory.go:223] Registration of the systemd container factory successfully Apr 14 13:31:50.380739 kubelet[2506]: I0414 13:31:50.380699 2506 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 14 13:31:50.382293 kubelet[2506]: I0414 13:31:50.382280 2506 factory.go:223] Registration of the containerd container factory successfully Apr 14 13:31:50.448643 kubelet[2506]: E0414 13:31:50.444481 2506 kubelet.go:1656] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 14 13:31:50.473458 kubelet[2506]: I0414 13:31:50.473149 2506 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Apr 14 13:31:50.481573 kubelet[2506]: I0414 13:31:50.481067 2506 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Apr 14 13:31:50.481573 kubelet[2506]: I0414 13:31:50.481091 2506 status_manager.go:249] "Starting to sync pod status with apiserver" Apr 14 13:31:50.481573 kubelet[2506]: I0414 13:31:50.481130 2506 kubelet.go:2501] "Starting kubelet main sync loop" Apr 14 13:31:50.481573 kubelet[2506]: E0414 13:31:50.481179 2506 kubelet.go:2525] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 14 13:31:50.531746 kubelet[2506]: I0414 13:31:50.531717 2506 cpu_manager.go:225] "Starting" policy="none" Apr 14 13:31:50.531941 kubelet[2506]: I0414 13:31:50.531931 2506 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Apr 14 13:31:50.533182 kubelet[2506]: I0414 13:31:50.532472 2506 state_mem.go:41] "Initialized" logger="CPUManager state checkpoint.CPUManager state memory" Apr 14 13:31:50.533182 kubelet[2506]: I0414 13:31:50.532581 2506 state_mem.go:94] "Updated default CPUSet" logger="CPUManager state checkpoint.CPUManager state memory" cpuSet="" Apr 14 13:31:50.533182 kubelet[2506]: I0414 13:31:50.532589 2506 state_mem.go:102] "Updated CPUSet assignments" logger="CPUManager state checkpoint.CPUManager state memory" assignments={} Apr 14 13:31:50.533182 kubelet[2506]: I0414 13:31:50.532603 2506 policy_none.go:50] "Start" Apr 14 13:31:50.533182 kubelet[2506]: I0414 13:31:50.532630 2506 memory_manager.go:187] "Starting memorymanager" policy="None" Apr 14 13:31:50.533182 kubelet[2506]: I0414 13:31:50.532646 2506 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Apr 14 13:31:50.533182 kubelet[2506]: I0414 13:31:50.532726 2506 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Apr 14 13:31:50.533182 kubelet[2506]: I0414 13:31:50.532732 2506 policy_none.go:44] "Start" Apr 14 13:31:50.540252 kubelet[2506]: E0414 13:31:50.540168 2506 manager.go:525] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 14 13:31:50.540566 kubelet[2506]: I0414 13:31:50.540542 2506 eviction_manager.go:194] "Eviction manager: starting control loop" Apr 14 13:31:50.540630 kubelet[2506]: I0414 13:31:50.540562 2506 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 14 13:31:50.542379 kubelet[2506]: I0414 13:31:50.540901 2506 plugin_manager.go:121] "Starting Kubelet Plugin Manager" Apr 14 13:31:50.544317 kubelet[2506]: E0414 13:31:50.544283 2506 eviction_manager.go:272] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 14 13:31:50.583750 kubelet[2506]: I0414 13:31:50.583393 2506 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Apr 14 13:31:50.642425 kubelet[2506]: I0414 13:31:50.590808 2506 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:50.642425 kubelet[2506]: I0414 13:31:50.592033 2506 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:50.728704 kubelet[2506]: E0414 13:31:50.728544 2506 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:50.782331 kubelet[2506]: I0414 13:31:50.781021 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3566c1d7ed03bb3c60facf009a5678dd-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"3566c1d7ed03bb3c60facf009a5678dd\") " pod="kube-system/kube-scheduler-localhost" Apr 14 13:31:50.782331 kubelet[2506]: I0414 13:31:50.781748 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/eb404586bd2117467da4e586f8d427df-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"eb404586bd2117467da4e586f8d427df\") " pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:50.782331 kubelet[2506]: I0414 13:31:50.781824 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/eb404586bd2117467da4e586f8d427df-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"eb404586bd2117467da4e586f8d427df\") " pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:50.782331 kubelet[2506]: I0414 13:31:50.781887 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:50.782331 kubelet[2506]: I0414 13:31:50.781917 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:50.785128 kubelet[2506]: I0414 13:31:50.781943 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:50.785128 kubelet[2506]: I0414 13:31:50.781961 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:50.785128 kubelet[2506]: I0414 13:31:50.781978 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/eb404586bd2117467da4e586f8d427df-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"eb404586bd2117467da4e586f8d427df\") " pod="kube-system/kube-apiserver-localhost" Apr 14 13:31:50.785128 kubelet[2506]: I0414 13:31:50.782010 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bd70d524e6bc561f2082b467706799ed-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"bd70d524e6bc561f2082b467706799ed\") " pod="kube-system/kube-controller-manager-localhost" Apr 14 13:31:50.792398 kubelet[2506]: E0414 13:31:50.792358 2506 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Apr 14 13:31:50.889848 kubelet[2506]: I0414 13:31:50.888993 2506 kubelet_node_status.go:74] "Attempting to register node" node="localhost" Apr 14 13:31:50.920044 kubelet[2506]: I0414 13:31:50.919953 2506 kubelet_node_status.go:123] "Node was previously registered" node="localhost" Apr 14 13:31:50.920044 kubelet[2506]: I0414 13:31:50.920062 2506 kubelet_node_status.go:77] "Successfully registered node" node="localhost" Apr 14 13:31:51.054954 kubelet[2506]: E0414 13:31:51.054861 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:51.093859 kubelet[2506]: E0414 13:31:51.093506 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:51.095462 kubelet[2506]: E0414 13:31:51.095322 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:51.337955 kubelet[2506]: I0414 13:31:51.337862 2506 apiserver.go:52] "Watching apiserver" Apr 14 13:31:51.376992 kubelet[2506]: I0414 13:31:51.376703 2506 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 14 13:31:51.560357 kubelet[2506]: E0414 13:31:51.560308 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:51.560526 kubelet[2506]: E0414 13:31:51.560426 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:51.561113 kubelet[2506]: E0414 13:31:51.561080 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:51.622177 kubelet[2506]: I0414 13:31:51.621898 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=3.6218716520000003 podStartE2EDuration="3.621871652s" podCreationTimestamp="2026-04-14 13:31:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-14 13:31:51.621675625 +0000 UTC m=+1.367682016" watchObservedRunningTime="2026-04-14 13:31:51.621871652 +0000 UTC m=+1.367878032" Apr 14 13:31:51.706172 kubelet[2506]: I0414 13:31:51.705048 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.705029982 podStartE2EDuration="1.705029982s" podCreationTimestamp="2026-04-14 13:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-14 13:31:51.676580983 +0000 UTC m=+1.422587371" watchObservedRunningTime="2026-04-14 13:31:51.705029982 +0000 UTC m=+1.451036377" Apr 14 13:31:52.564023 kubelet[2506]: E0414 13:31:52.562539 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:52.564023 kubelet[2506]: E0414 13:31:52.562626 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:53.763674 kubelet[2506]: I0414 13:31:53.763624 2506 kuberuntime_manager.go:2062] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 14 13:31:53.764325 kubelet[2506]: I0414 13:31:53.764311 2506 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 14 13:31:53.764372 containerd[1454]: time="2026-04-14T13:31:53.764088969Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 14 13:31:54.377981 systemd[1]: Created slice kubepods-besteffort-podf9f9d922_d7ee_465a_8d3e_eb9bffd26675.slice - libcontainer container kubepods-besteffort-podf9f9d922_d7ee_465a_8d3e_eb9bffd26675.slice. Apr 14 13:31:54.527068 kubelet[2506]: I0414 13:31:54.526952 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/f9f9d922-d7ee-465a-8d3e-eb9bffd26675-kube-proxy\") pod \"kube-proxy-q5mjv\" (UID: \"f9f9d922-d7ee-465a-8d3e-eb9bffd26675\") " pod="kube-system/kube-proxy-q5mjv" Apr 14 13:31:54.527068 kubelet[2506]: I0414 13:31:54.527030 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f9f9d922-d7ee-465a-8d3e-eb9bffd26675-xtables-lock\") pod \"kube-proxy-q5mjv\" (UID: \"f9f9d922-d7ee-465a-8d3e-eb9bffd26675\") " pod="kube-system/kube-proxy-q5mjv" Apr 14 13:31:54.527068 kubelet[2506]: I0414 13:31:54.527055 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9f9d922-d7ee-465a-8d3e-eb9bffd26675-lib-modules\") pod \"kube-proxy-q5mjv\" (UID: \"f9f9d922-d7ee-465a-8d3e-eb9bffd26675\") " pod="kube-system/kube-proxy-q5mjv" Apr 14 13:31:54.527068 kubelet[2506]: I0414 13:31:54.527069 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8q2v\" (UniqueName: \"kubernetes.io/projected/f9f9d922-d7ee-465a-8d3e-eb9bffd26675-kube-api-access-p8q2v\") pod \"kube-proxy-q5mjv\" (UID: \"f9f9d922-d7ee-465a-8d3e-eb9bffd26675\") " pod="kube-system/kube-proxy-q5mjv" Apr 14 13:31:54.580101 systemd[1]: Created slice kubepods-besteffort-poda15a9384_10b5_4088_bada_607344052ec6.slice - libcontainer container kubepods-besteffort-poda15a9384_10b5_4088_bada_607344052ec6.slice. Apr 14 13:31:54.689245 kubelet[2506]: E0414 13:31:54.689162 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:54.690103 containerd[1454]: time="2026-04-14T13:31:54.690050385Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-q5mjv,Uid:f9f9d922-d7ee-465a-8d3e-eb9bffd26675,Namespace:kube-system,Attempt:0,}" Apr 14 13:31:54.732281 kubelet[2506]: I0414 13:31:54.731164 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/a15a9384-10b5-4088-bada-607344052ec6-var-lib-calico\") pod \"tigera-operator-6cf4cccc57-4n9gm\" (UID: \"a15a9384-10b5-4088-bada-607344052ec6\") " pod="tigera-operator/tigera-operator-6cf4cccc57-4n9gm" Apr 14 13:31:54.732281 kubelet[2506]: I0414 13:31:54.731273 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l72j\" (UniqueName: \"kubernetes.io/projected/a15a9384-10b5-4088-bada-607344052ec6-kube-api-access-8l72j\") pod \"tigera-operator-6cf4cccc57-4n9gm\" (UID: \"a15a9384-10b5-4088-bada-607344052ec6\") " pod="tigera-operator/tigera-operator-6cf4cccc57-4n9gm" Apr 14 13:31:54.739193 containerd[1454]: time="2026-04-14T13:31:54.737010023Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:31:54.739193 containerd[1454]: time="2026-04-14T13:31:54.737610575Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:31:54.739193 containerd[1454]: time="2026-04-14T13:31:54.737621709Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:54.739557 containerd[1454]: time="2026-04-14T13:31:54.739087573Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:54.782892 systemd[1]: Started cri-containerd-64a0100a98f69cea2c4dd0b8388af15b9dead9ea37e8afecc9de7ff0f18ba481.scope - libcontainer container 64a0100a98f69cea2c4dd0b8388af15b9dead9ea37e8afecc9de7ff0f18ba481. Apr 14 13:31:55.168316 containerd[1454]: time="2026-04-14T13:31:55.168264400Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-q5mjv,Uid:f9f9d922-d7ee-465a-8d3e-eb9bffd26675,Namespace:kube-system,Attempt:0,} returns sandbox id \"64a0100a98f69cea2c4dd0b8388af15b9dead9ea37e8afecc9de7ff0f18ba481\"" Apr 14 13:31:55.171441 kubelet[2506]: E0414 13:31:55.171382 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:55.180166 containerd[1454]: time="2026-04-14T13:31:55.180086648Z" level=info msg="CreateContainer within sandbox \"64a0100a98f69cea2c4dd0b8388af15b9dead9ea37e8afecc9de7ff0f18ba481\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 14 13:31:55.199461 containerd[1454]: time="2026-04-14T13:31:55.199370541Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6cf4cccc57-4n9gm,Uid:a15a9384-10b5-4088-bada-607344052ec6,Namespace:tigera-operator,Attempt:0,}" Apr 14 13:31:55.232826 containerd[1454]: time="2026-04-14T13:31:55.232367488Z" level=info msg="CreateContainer within sandbox \"64a0100a98f69cea2c4dd0b8388af15b9dead9ea37e8afecc9de7ff0f18ba481\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4fcba6f7c5d295b67d5c1e9cd079b31ba3391e8b3f0c181e6b16fe19f0c035ae\"" Apr 14 13:31:55.234933 containerd[1454]: time="2026-04-14T13:31:55.234727089Z" level=info msg="StartContainer for \"4fcba6f7c5d295b67d5c1e9cd079b31ba3391e8b3f0c181e6b16fe19f0c035ae\"" Apr 14 13:31:55.314572 containerd[1454]: time="2026-04-14T13:31:55.313446707Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:31:55.316337 containerd[1454]: time="2026-04-14T13:31:55.315875306Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:31:55.317720 containerd[1454]: time="2026-04-14T13:31:55.316696670Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:55.317720 containerd[1454]: time="2026-04-14T13:31:55.317315065Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:31:55.343441 systemd[1]: Started cri-containerd-4fcba6f7c5d295b67d5c1e9cd079b31ba3391e8b3f0c181e6b16fe19f0c035ae.scope - libcontainer container 4fcba6f7c5d295b67d5c1e9cd079b31ba3391e8b3f0c181e6b16fe19f0c035ae. Apr 14 13:31:55.359110 systemd[1]: Started cri-containerd-4855324da7e6612f5a7ede1e9ba4d8b86ccb8c5126977ff53f98f6abd8b52c71.scope - libcontainer container 4855324da7e6612f5a7ede1e9ba4d8b86ccb8c5126977ff53f98f6abd8b52c71. Apr 14 13:31:55.478668 containerd[1454]: time="2026-04-14T13:31:55.478315364Z" level=info msg="StartContainer for \"4fcba6f7c5d295b67d5c1e9cd079b31ba3391e8b3f0c181e6b16fe19f0c035ae\" returns successfully" Apr 14 13:31:55.515430 containerd[1454]: time="2026-04-14T13:31:55.515199427Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6cf4cccc57-4n9gm,Uid:a15a9384-10b5-4088-bada-607344052ec6,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"4855324da7e6612f5a7ede1e9ba4d8b86ccb8c5126977ff53f98f6abd8b52c71\"" Apr 14 13:31:55.528374 containerd[1454]: time="2026-04-14T13:31:55.528277969Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\"" Apr 14 13:31:55.636525 kubelet[2506]: E0414 13:31:55.634209 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:31:57.597560 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1964385589.mount: Deactivated successfully. Apr 14 13:31:58.835389 containerd[1454]: time="2026-04-14T13:31:58.835333368Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:58.836271 containerd[1454]: time="2026-04-14T13:31:58.836237850Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.7: active requests=0, bytes read=40846156" Apr 14 13:31:58.839609 containerd[1454]: time="2026-04-14T13:31:58.839532786Z" level=info msg="ImageCreate event name:\"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:58.842746 containerd[1454]: time="2026-04-14T13:31:58.842706586Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:31:58.843452 containerd[1454]: time="2026-04-14T13:31:58.843417363Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.7\" with image id \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\", repo tag \"quay.io/tigera/operator:v1.40.7\", repo digest \"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\", size \"40842151\" in 3.314002014s" Apr 14 13:31:58.843486 containerd[1454]: time="2026-04-14T13:31:58.843458517Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\" returns image reference \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\"" Apr 14 13:31:58.850404 containerd[1454]: time="2026-04-14T13:31:58.850361963Z" level=info msg="CreateContainer within sandbox \"4855324da7e6612f5a7ede1e9ba4d8b86ccb8c5126977ff53f98f6abd8b52c71\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Apr 14 13:31:58.870291 containerd[1454]: time="2026-04-14T13:31:58.870018603Z" level=info msg="CreateContainer within sandbox \"4855324da7e6612f5a7ede1e9ba4d8b86ccb8c5126977ff53f98f6abd8b52c71\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"c5ca7eefa53af1eed4787c753d2416ac2d5d8bc19f83155cfc8bce9a1a161f77\"" Apr 14 13:31:58.872202 containerd[1454]: time="2026-04-14T13:31:58.872123495Z" level=info msg="StartContainer for \"c5ca7eefa53af1eed4787c753d2416ac2d5d8bc19f83155cfc8bce9a1a161f77\"" Apr 14 13:31:58.921478 systemd[1]: Started cri-containerd-c5ca7eefa53af1eed4787c753d2416ac2d5d8bc19f83155cfc8bce9a1a161f77.scope - libcontainer container c5ca7eefa53af1eed4787c753d2416ac2d5d8bc19f83155cfc8bce9a1a161f77. Apr 14 13:31:58.951711 containerd[1454]: time="2026-04-14T13:31:58.951584563Z" level=info msg="StartContainer for \"c5ca7eefa53af1eed4787c753d2416ac2d5d8bc19f83155cfc8bce9a1a161f77\" returns successfully" Apr 14 13:31:59.778869 kubelet[2506]: I0414 13:31:59.778592 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-proxy-q5mjv" podStartSLOduration=5.778576246 podStartE2EDuration="5.778576246s" podCreationTimestamp="2026-04-14 13:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-14 13:31:55.700972701 +0000 UTC m=+5.446979089" watchObservedRunningTime="2026-04-14 13:31:59.778576246 +0000 UTC m=+9.524582667" Apr 14 13:32:00.516795 kubelet[2506]: I0414 13:32:00.516713 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6cf4cccc57-4n9gm" podStartSLOduration=3.197905383 podStartE2EDuration="6.516703609s" podCreationTimestamp="2026-04-14 13:31:54 +0000 UTC" firstStartedPulling="2026-04-14 13:31:55.526642661 +0000 UTC m=+5.272649041" lastFinishedPulling="2026-04-14 13:31:58.845440887 +0000 UTC m=+8.591447267" observedRunningTime="2026-04-14 13:31:59.781731483 +0000 UTC m=+9.527737874" watchObservedRunningTime="2026-04-14 13:32:00.516703609 +0000 UTC m=+10.262710000" Apr 14 13:32:04.716306 sudo[1632]: pam_unix(sudo:session): session closed for user root Apr 14 13:32:04.719700 sshd[1629]: pam_unix(sshd:session): session closed for user core Apr 14 13:32:04.728076 systemd[1]: sshd@6-10.0.0.8:22-10.0.0.1:42804.service: Deactivated successfully. Apr 14 13:32:04.737546 systemd[1]: session-7.scope: Deactivated successfully. Apr 14 13:32:04.740066 systemd[1]: session-7.scope: Consumed 5.920s CPU time, 158.2M memory peak, 0B memory swap peak. Apr 14 13:32:04.742376 systemd-logind[1442]: Session 7 logged out. Waiting for processes to exit. Apr 14 13:32:04.750103 systemd-logind[1442]: Removed session 7. Apr 14 13:32:04.805331 update_engine[1445]: I20260414 13:32:04.805054 1445 update_attempter.cc:509] Updating boot flags... Apr 14 13:32:04.882289 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (2924) Apr 14 13:32:10.456081 systemd[1]: Created slice kubepods-besteffort-pod30dd40e2_34af_4d02_ab75_c1f4d0654047.slice - libcontainer container kubepods-besteffort-pod30dd40e2_34af_4d02_ab75_c1f4d0654047.slice. Apr 14 13:32:10.645911 kubelet[2506]: I0414 13:32:10.645626 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-bpffs\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.645911 kubelet[2506]: I0414 13:32:10.645877 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-cni-net-dir\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.646785 kubelet[2506]: I0414 13:32:10.645945 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-lib-modules\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.646785 kubelet[2506]: I0414 13:32:10.645986 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpnxz\" (UniqueName: \"kubernetes.io/projected/30dd40e2-34af-4d02-ab75-c1f4d0654047-kube-api-access-wpnxz\") pod \"calico-typha-6847b996d8-mh4vw\" (UID: \"30dd40e2-34af-4d02-ab75-c1f4d0654047\") " pod="calico-system/calico-typha-6847b996d8-mh4vw" Apr 14 13:32:10.646785 kubelet[2506]: I0414 13:32:10.646004 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-cni-log-dir\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.646785 kubelet[2506]: I0414 13:32:10.646042 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/30dd40e2-34af-4d02-ab75-c1f4d0654047-typha-certs\") pod \"calico-typha-6847b996d8-mh4vw\" (UID: \"30dd40e2-34af-4d02-ab75-c1f4d0654047\") " pod="calico-system/calico-typha-6847b996d8-mh4vw" Apr 14 13:32:10.646785 kubelet[2506]: I0414 13:32:10.646058 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30dd40e2-34af-4d02-ab75-c1f4d0654047-tigera-ca-bundle\") pod \"calico-typha-6847b996d8-mh4vw\" (UID: \"30dd40e2-34af-4d02-ab75-c1f4d0654047\") " pod="calico-system/calico-typha-6847b996d8-mh4vw" Apr 14 13:32:10.648994 kubelet[2506]: I0414 13:32:10.646075 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-flexvol-driver-host\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.648994 kubelet[2506]: I0414 13:32:10.646094 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-sys-fs\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.648994 kubelet[2506]: I0414 13:32:10.646108 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/f458aee8-675f-406a-b973-ea55edc77b9d-node-certs\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.648994 kubelet[2506]: I0414 13:32:10.646119 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f458aee8-675f-406a-b973-ea55edc77b9d-tigera-ca-bundle\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.648994 kubelet[2506]: I0414 13:32:10.646195 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh9f2\" (UniqueName: \"kubernetes.io/projected/f458aee8-675f-406a-b973-ea55edc77b9d-kube-api-access-fh9f2\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.649203 kubelet[2506]: I0414 13:32:10.646209 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-policysync\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.649203 kubelet[2506]: I0414 13:32:10.646248 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-nodeproc\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.649203 kubelet[2506]: I0414 13:32:10.646260 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-xtables-lock\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.649203 kubelet[2506]: I0414 13:32:10.646273 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-cni-bin-dir\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.649203 kubelet[2506]: I0414 13:32:10.647743 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-var-lib-calico\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.649417 kubelet[2506]: I0414 13:32:10.647901 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/f458aee8-675f-406a-b973-ea55edc77b9d-var-run-calico\") pod \"calico-node-ch7dd\" (UID: \"f458aee8-675f-406a-b973-ea55edc77b9d\") " pod="calico-system/calico-node-ch7dd" Apr 14 13:32:10.658667 systemd[1]: Created slice kubepods-besteffort-podf458aee8_675f_406a_b973_ea55edc77b9d.slice - libcontainer container kubepods-besteffort-podf458aee8_675f_406a_b973_ea55edc77b9d.slice. Apr 14 13:32:10.773968 kubelet[2506]: E0414 13:32:10.773661 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.773968 kubelet[2506]: W0414 13:32:10.773689 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.773968 kubelet[2506]: E0414 13:32:10.773750 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.782759 kubelet[2506]: E0414 13:32:10.781068 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.790165 kubelet[2506]: W0414 13:32:10.789799 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.790165 kubelet[2506]: E0414 13:32:10.789967 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.793355 kubelet[2506]: E0414 13:32:10.791017 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.793355 kubelet[2506]: W0414 13:32:10.791037 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.793355 kubelet[2506]: E0414 13:32:10.791055 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.797192 kubelet[2506]: E0414 13:32:10.797171 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.797389 kubelet[2506]: W0414 13:32:10.797373 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.797483 kubelet[2506]: E0414 13:32:10.797468 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.797836 kubelet[2506]: E0414 13:32:10.797824 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.797925 kubelet[2506]: W0414 13:32:10.797915 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.797986 kubelet[2506]: E0414 13:32:10.797976 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.798357 kubelet[2506]: E0414 13:32:10.798344 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.798510 kubelet[2506]: W0414 13:32:10.798497 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.802713 kubelet[2506]: E0414 13:32:10.802522 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.803816 kubelet[2506]: E0414 13:32:10.803755 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.804029 kubelet[2506]: W0414 13:32:10.803960 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.804141 kubelet[2506]: E0414 13:32:10.804130 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.804641 kubelet[2506]: E0414 13:32:10.804630 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.804759 kubelet[2506]: W0414 13:32:10.804746 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.804816 kubelet[2506]: E0414 13:32:10.804807 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.805126 kubelet[2506]: E0414 13:32:10.805115 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.805200 kubelet[2506]: W0414 13:32:10.805189 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.807270 kubelet[2506]: E0414 13:32:10.807096 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.819532 kubelet[2506]: E0414 13:32:10.817651 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.819532 kubelet[2506]: W0414 13:32:10.817959 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.819532 kubelet[2506]: E0414 13:32:10.818113 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.822309 kubelet[2506]: E0414 13:32:10.822279 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.824215 kubelet[2506]: W0414 13:32:10.824192 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.824878 kubelet[2506]: E0414 13:32:10.824798 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.826087 kubelet[2506]: E0414 13:32:10.825328 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.826087 kubelet[2506]: W0414 13:32:10.825342 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.826087 kubelet[2506]: E0414 13:32:10.825357 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.834522 kubelet[2506]: E0414 13:32:10.834403 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.835207 kubelet[2506]: W0414 13:32:10.835179 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.835374 kubelet[2506]: E0414 13:32:10.835357 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.835949 kubelet[2506]: E0414 13:32:10.835935 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.836020 kubelet[2506]: W0414 13:32:10.836009 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.836393 kubelet[2506]: E0414 13:32:10.836296 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.838622 kubelet[2506]: E0414 13:32:10.838199 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.855064 kubelet[2506]: W0414 13:32:10.850011 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.855064 kubelet[2506]: E0414 13:32:10.850199 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.855064 kubelet[2506]: E0414 13:32:10.851088 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.855064 kubelet[2506]: W0414 13:32:10.851098 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.855064 kubelet[2506]: E0414 13:32:10.851174 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.855064 kubelet[2506]: E0414 13:32:10.851415 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.855064 kubelet[2506]: W0414 13:32:10.851422 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.855064 kubelet[2506]: E0414 13:32:10.851430 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.855064 kubelet[2506]: E0414 13:32:10.851610 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.855064 kubelet[2506]: W0414 13:32:10.851615 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.855438 kubelet[2506]: E0414 13:32:10.851622 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.857119 kubelet[2506]: E0414 13:32:10.855479 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.857119 kubelet[2506]: W0414 13:32:10.855522 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.857119 kubelet[2506]: E0414 13:32:10.855536 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.857119 kubelet[2506]: E0414 13:32:10.855819 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.857119 kubelet[2506]: W0414 13:32:10.855828 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.857119 kubelet[2506]: E0414 13:32:10.855840 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.859772 kubelet[2506]: E0414 13:32:10.859671 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.859772 kubelet[2506]: W0414 13:32:10.859744 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.859772 kubelet[2506]: E0414 13:32:10.859765 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.860544 kubelet[2506]: E0414 13:32:10.860382 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.860544 kubelet[2506]: W0414 13:32:10.860397 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.860544 kubelet[2506]: E0414 13:32:10.860412 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.862499 kubelet[2506]: E0414 13:32:10.862388 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.862499 kubelet[2506]: W0414 13:32:10.862416 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.864001 kubelet[2506]: E0414 13:32:10.863171 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.877639 kubelet[2506]: E0414 13:32:10.876863 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.877639 kubelet[2506]: W0414 13:32:10.876966 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.877639 kubelet[2506]: E0414 13:32:10.877031 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.880887 kubelet[2506]: E0414 13:32:10.879432 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.880887 kubelet[2506]: W0414 13:32:10.879505 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.880887 kubelet[2506]: E0414 13:32:10.879524 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.883510 kubelet[2506]: E0414 13:32:10.883103 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.887977 kubelet[2506]: W0414 13:32:10.885131 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.887977 kubelet[2506]: E0414 13:32:10.886396 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.895272 kubelet[2506]: E0414 13:32:10.895040 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.904779 kubelet[2506]: W0414 13:32:10.895205 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.957755 kubelet[2506]: E0414 13:32:10.957714 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.959170 kubelet[2506]: E0414 13:32:10.959080 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.959283 kubelet[2506]: W0414 13:32:10.959161 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.960271 kubelet[2506]: E0414 13:32:10.959280 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.960764 kubelet[2506]: E0414 13:32:10.960421 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.962377 kubelet[2506]: W0414 13:32:10.961207 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.962478 kubelet[2506]: E0414 13:32:10.962382 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.966768 kubelet[2506]: E0414 13:32:10.966686 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.966888 kubelet[2506]: W0414 13:32:10.966758 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.966980 kubelet[2506]: E0414 13:32:10.966907 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.968259 kubelet[2506]: E0414 13:32:10.968194 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.968305 kubelet[2506]: W0414 13:32:10.968293 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.968326 kubelet[2506]: E0414 13:32:10.968311 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.978200 kubelet[2506]: E0414 13:32:10.977961 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.978448 kubelet[2506]: W0414 13:32:10.978191 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.978448 kubelet[2506]: E0414 13:32:10.978386 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:10.986362 kubelet[2506]: E0414 13:32:10.984146 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:10.986814 kubelet[2506]: W0414 13:32:10.986412 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:10.986814 kubelet[2506]: E0414 13:32:10.986579 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.004193 kubelet[2506]: E0414 13:32:11.002055 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.004672 kubelet[2506]: W0414 13:32:11.004529 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.013091 kubelet[2506]: E0414 13:32:11.010633 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.019315 kubelet[2506]: E0414 13:32:11.019191 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.019544 kubelet[2506]: W0414 13:32:11.019300 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.019544 kubelet[2506]: E0414 13:32:11.019452 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.020147 kubelet[2506]: E0414 13:32:11.020075 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.020147 kubelet[2506]: W0414 13:32:11.020104 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.020147 kubelet[2506]: E0414 13:32:11.020118 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.020494 kubelet[2506]: E0414 13:32:11.020444 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.020694 kubelet[2506]: W0414 13:32:11.020528 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.020694 kubelet[2506]: E0414 13:32:11.020542 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.021498 kubelet[2506]: E0414 13:32:11.021447 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.021498 kubelet[2506]: W0414 13:32:11.021473 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.021498 kubelet[2506]: E0414 13:32:11.021484 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.070477 kubelet[2506]: E0414 13:32:11.070157 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:11.093927 kubelet[2506]: E0414 13:32:11.093126 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:32:11.102558 containerd[1454]: time="2026-04-14T13:32:11.102484932Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6847b996d8-mh4vw,Uid:30dd40e2-34af-4d02-ab75-c1f4d0654047,Namespace:calico-system,Attempt:0,}" Apr 14 13:32:11.179397 kubelet[2506]: E0414 13:32:11.179319 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.179397 kubelet[2506]: W0414 13:32:11.179374 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.179397 kubelet[2506]: E0414 13:32:11.179422 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.190302 kubelet[2506]: E0414 13:32:11.189484 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.190302 kubelet[2506]: W0414 13:32:11.189638 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.190302 kubelet[2506]: E0414 13:32:11.189761 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.247933 kubelet[2506]: E0414 13:32:11.243882 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.247933 kubelet[2506]: W0414 13:32:11.243946 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.247933 kubelet[2506]: E0414 13:32:11.243969 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.247933 kubelet[2506]: E0414 13:32:11.244536 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.247933 kubelet[2506]: W0414 13:32:11.244548 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.247933 kubelet[2506]: E0414 13:32:11.244563 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.250295 kubelet[2506]: E0414 13:32:11.248801 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.250295 kubelet[2506]: W0414 13:32:11.248991 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.250295 kubelet[2506]: E0414 13:32:11.249279 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.250295 kubelet[2506]: E0414 13:32:11.250168 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.250295 kubelet[2506]: W0414 13:32:11.250178 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.250295 kubelet[2506]: E0414 13:32:11.250192 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.250946 kubelet[2506]: E0414 13:32:11.250870 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.250946 kubelet[2506]: W0414 13:32:11.250900 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.250946 kubelet[2506]: E0414 13:32:11.250913 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.251065 kubelet[2506]: E0414 13:32:11.251054 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.251065 kubelet[2506]: W0414 13:32:11.251060 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.251126 kubelet[2506]: E0414 13:32:11.251068 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.251528 kubelet[2506]: E0414 13:32:11.251200 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.251528 kubelet[2506]: W0414 13:32:11.251209 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.251528 kubelet[2506]: E0414 13:32:11.251258 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.251528 kubelet[2506]: E0414 13:32:11.251459 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.251528 kubelet[2506]: W0414 13:32:11.251468 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.251528 kubelet[2506]: E0414 13:32:11.251477 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.251750 kubelet[2506]: E0414 13:32:11.251716 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.251750 kubelet[2506]: W0414 13:32:11.251724 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.251750 kubelet[2506]: E0414 13:32:11.251733 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.254392 kubelet[2506]: E0414 13:32:11.251913 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.254392 kubelet[2506]: W0414 13:32:11.251924 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.254392 kubelet[2506]: E0414 13:32:11.251934 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.254392 kubelet[2506]: E0414 13:32:11.252137 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.254392 kubelet[2506]: W0414 13:32:11.252161 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.254392 kubelet[2506]: E0414 13:32:11.252171 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.259181 kubelet[2506]: E0414 13:32:11.258671 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.260969 kubelet[2506]: W0414 13:32:11.260275 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.260969 kubelet[2506]: E0414 13:32:11.260570 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.261914 kubelet[2506]: E0414 13:32:11.261713 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.263407 kubelet[2506]: W0414 13:32:11.263281 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.268644 kubelet[2506]: E0414 13:32:11.268200 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.270264 kubelet[2506]: E0414 13:32:11.269277 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.270264 kubelet[2506]: W0414 13:32:11.269292 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.270264 kubelet[2506]: E0414 13:32:11.269528 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.286098 containerd[1454]: time="2026-04-14T13:32:11.285157542Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:32:11.287199 containerd[1454]: time="2026-04-14T13:32:11.285848838Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:32:11.287514 containerd[1454]: time="2026-04-14T13:32:11.285871370Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:32:11.295866 containerd[1454]: time="2026-04-14T13:32:11.292132496Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:32:11.300175 kubelet[2506]: E0414 13:32:11.284182 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.300363 kubelet[2506]: W0414 13:32:11.300245 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.300397 kubelet[2506]: E0414 13:32:11.300384 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.309593 kubelet[2506]: E0414 13:32:11.308049 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.309593 kubelet[2506]: W0414 13:32:11.308071 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.309593 kubelet[2506]: E0414 13:32:11.308091 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.313319 kubelet[2506]: E0414 13:32:11.312058 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.313319 kubelet[2506]: W0414 13:32:11.312905 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.313319 kubelet[2506]: E0414 13:32:11.313194 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.313922 kubelet[2506]: E0414 13:32:11.313909 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.313989 kubelet[2506]: W0414 13:32:11.313979 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.314495 kubelet[2506]: E0414 13:32:11.314034 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.324863 kubelet[2506]: E0414 13:32:11.320706 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.324863 kubelet[2506]: W0414 13:32:11.324361 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.332110 kubelet[2506]: E0414 13:32:11.329665 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.332352 kubelet[2506]: I0414 13:32:11.332311 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/b5b0d33b-1b54-43cc-a4d3-cd788e8eb914-varrun\") pod \"csi-node-driver-5xvdj\" (UID: \"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914\") " pod="calico-system/csi-node-driver-5xvdj" Apr 14 13:32:11.334794 containerd[1454]: time="2026-04-14T13:32:11.334751612Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ch7dd,Uid:f458aee8-675f-406a-b973-ea55edc77b9d,Namespace:calico-system,Attempt:0,}" Apr 14 13:32:11.335449 kubelet[2506]: E0414 13:32:11.335198 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.335449 kubelet[2506]: W0414 13:32:11.335281 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.335449 kubelet[2506]: E0414 13:32:11.335302 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.336037 kubelet[2506]: E0414 13:32:11.336015 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.336117 kubelet[2506]: W0414 13:32:11.336029 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.336117 kubelet[2506]: E0414 13:32:11.336071 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.349823 kubelet[2506]: E0414 13:32:11.345852 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.350526 kubelet[2506]: W0414 13:32:11.350181 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.350721 kubelet[2506]: E0414 13:32:11.350631 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.351043 kubelet[2506]: I0414 13:32:11.350994 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5b0d33b-1b54-43cc-a4d3-cd788e8eb914-kubelet-dir\") pod \"csi-node-driver-5xvdj\" (UID: \"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914\") " pod="calico-system/csi-node-driver-5xvdj" Apr 14 13:32:11.351328 kubelet[2506]: E0414 13:32:11.351265 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.351328 kubelet[2506]: W0414 13:32:11.351292 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.351398 kubelet[2506]: E0414 13:32:11.351336 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.351623 kubelet[2506]: E0414 13:32:11.351579 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.351623 kubelet[2506]: W0414 13:32:11.351614 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.351713 kubelet[2506]: E0414 13:32:11.351626 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.351914 kubelet[2506]: E0414 13:32:11.351874 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.351914 kubelet[2506]: W0414 13:32:11.351908 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.351980 kubelet[2506]: E0414 13:32:11.351919 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.351980 kubelet[2506]: I0414 13:32:11.351961 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b5b0d33b-1b54-43cc-a4d3-cd788e8eb914-socket-dir\") pod \"csi-node-driver-5xvdj\" (UID: \"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914\") " pod="calico-system/csi-node-driver-5xvdj" Apr 14 13:32:11.352496 kubelet[2506]: E0414 13:32:11.352440 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.352555 kubelet[2506]: W0414 13:32:11.352496 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.352555 kubelet[2506]: E0414 13:32:11.352513 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.352555 kubelet[2506]: I0414 13:32:11.352533 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b5b0d33b-1b54-43cc-a4d3-cd788e8eb914-registration-dir\") pod \"csi-node-driver-5xvdj\" (UID: \"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914\") " pod="calico-system/csi-node-driver-5xvdj" Apr 14 13:32:11.370032 kubelet[2506]: E0414 13:32:11.369613 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.371208 kubelet[2506]: W0414 13:32:11.369948 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.371208 kubelet[2506]: E0414 13:32:11.370131 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.371208 kubelet[2506]: I0414 13:32:11.370176 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7x9d\" (UniqueName: \"kubernetes.io/projected/b5b0d33b-1b54-43cc-a4d3-cd788e8eb914-kube-api-access-r7x9d\") pod \"csi-node-driver-5xvdj\" (UID: \"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914\") " pod="calico-system/csi-node-driver-5xvdj" Apr 14 13:32:11.376383 kubelet[2506]: E0414 13:32:11.375467 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.376383 kubelet[2506]: W0414 13:32:11.375529 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.376383 kubelet[2506]: E0414 13:32:11.375626 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.379988 kubelet[2506]: E0414 13:32:11.377538 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.379988 kubelet[2506]: W0414 13:32:11.377640 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.379538 systemd[1]: Started cri-containerd-9ecbdd7bf008eb04191e19cf678280ce120b63d7a524ede09b829a14283e529c.scope - libcontainer container 9ecbdd7bf008eb04191e19cf678280ce120b63d7a524ede09b829a14283e529c. Apr 14 13:32:11.382748 kubelet[2506]: E0414 13:32:11.382547 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.396590 kubelet[2506]: E0414 13:32:11.396563 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.396955 kubelet[2506]: W0414 13:32:11.396937 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.455343 kubelet[2506]: E0414 13:32:11.453841 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.473013 kubelet[2506]: E0414 13:32:11.472749 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.473013 kubelet[2506]: W0414 13:32:11.472970 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.473336 kubelet[2506]: E0414 13:32:11.473008 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.473844 kubelet[2506]: E0414 13:32:11.473792 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.473844 kubelet[2506]: W0414 13:32:11.473827 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.473844 kubelet[2506]: E0414 13:32:11.473843 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.475037 kubelet[2506]: E0414 13:32:11.474943 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.475037 kubelet[2506]: W0414 13:32:11.475021 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.475207 kubelet[2506]: E0414 13:32:11.475109 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.490663 kubelet[2506]: E0414 13:32:11.490591 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.490663 kubelet[2506]: W0414 13:32:11.490640 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.490663 kubelet[2506]: E0414 13:32:11.490664 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.495067 kubelet[2506]: E0414 13:32:11.493601 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.495067 kubelet[2506]: W0414 13:32:11.493678 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.495067 kubelet[2506]: E0414 13:32:11.494670 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.498937 kubelet[2506]: E0414 13:32:11.498305 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.498937 kubelet[2506]: W0414 13:32:11.498360 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.498937 kubelet[2506]: E0414 13:32:11.498428 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.499193 kubelet[2506]: E0414 13:32:11.499024 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.499193 kubelet[2506]: W0414 13:32:11.499034 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.499193 kubelet[2506]: E0414 13:32:11.499048 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.512370 kubelet[2506]: E0414 13:32:11.512092 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.512370 kubelet[2506]: W0414 13:32:11.512312 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.513048 kubelet[2506]: E0414 13:32:11.512754 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.522130 kubelet[2506]: E0414 13:32:11.522008 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.522130 kubelet[2506]: W0414 13:32:11.522099 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.522368 kubelet[2506]: E0414 13:32:11.522291 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.528285 kubelet[2506]: E0414 13:32:11.528099 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.528285 kubelet[2506]: W0414 13:32:11.528126 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.528285 kubelet[2506]: E0414 13:32:11.528154 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.535662 kubelet[2506]: E0414 13:32:11.535502 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.535662 kubelet[2506]: W0414 13:32:11.535542 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.535662 kubelet[2506]: E0414 13:32:11.535623 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.546580 kubelet[2506]: E0414 13:32:11.545034 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.546580 kubelet[2506]: W0414 13:32:11.545067 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.546580 kubelet[2506]: E0414 13:32:11.545148 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.564802 kubelet[2506]: E0414 13:32:11.564775 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.565264 kubelet[2506]: W0414 13:32:11.564985 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.565264 kubelet[2506]: E0414 13:32:11.565018 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.565531 kubelet[2506]: E0414 13:32:11.565520 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.565601 kubelet[2506]: W0414 13:32:11.565590 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.565657 kubelet[2506]: E0414 13:32:11.565648 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.565912 kubelet[2506]: E0414 13:32:11.565901 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.565978 kubelet[2506]: W0414 13:32:11.565967 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.566024 kubelet[2506]: E0414 13:32:11.566016 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.566211 kubelet[2506]: E0414 13:32:11.566201 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.566316 kubelet[2506]: W0414 13:32:11.566307 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.566410 kubelet[2506]: E0414 13:32:11.566357 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.566578 kubelet[2506]: E0414 13:32:11.566571 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.566610 kubelet[2506]: W0414 13:32:11.566605 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.566642 kubelet[2506]: E0414 13:32:11.566637 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.566850 kubelet[2506]: E0414 13:32:11.566843 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.566890 kubelet[2506]: W0414 13:32:11.566885 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.566963 kubelet[2506]: E0414 13:32:11.566914 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.567174 kubelet[2506]: E0414 13:32:11.567167 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.567278 kubelet[2506]: W0414 13:32:11.567212 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.567313 kubelet[2506]: E0414 13:32:11.567307 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.571889 kubelet[2506]: E0414 13:32:11.571873 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.572043 kubelet[2506]: W0414 13:32:11.571958 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.572043 kubelet[2506]: E0414 13:32:11.571976 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.572330 kubelet[2506]: E0414 13:32:11.572321 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.572396 kubelet[2506]: W0414 13:32:11.572389 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.572454 kubelet[2506]: E0414 13:32:11.572446 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.572813 kubelet[2506]: E0414 13:32:11.572712 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.572813 kubelet[2506]: W0414 13:32:11.572742 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.572813 kubelet[2506]: E0414 13:32:11.572754 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.573194 kubelet[2506]: E0414 13:32:11.573184 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.573341 kubelet[2506]: W0414 13:32:11.573285 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.573341 kubelet[2506]: E0414 13:32:11.573298 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.573559 kubelet[2506]: E0414 13:32:11.573553 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.573646 kubelet[2506]: W0414 13:32:11.573606 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.573646 kubelet[2506]: E0414 13:32:11.573616 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.578148 kubelet[2506]: E0414 13:32:11.575620 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.578148 kubelet[2506]: W0414 13:32:11.576182 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.578148 kubelet[2506]: E0414 13:32:11.576774 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.593360 kubelet[2506]: E0414 13:32:11.590773 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.595828 kubelet[2506]: W0414 13:32:11.594359 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.597321 kubelet[2506]: E0414 13:32:11.597144 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.597965 kubelet[2506]: E0414 13:32:11.597953 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.598027 kubelet[2506]: W0414 13:32:11.598018 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.598066 kubelet[2506]: E0414 13:32:11.598061 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.600496 kubelet[2506]: E0414 13:32:11.599670 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.602003 kubelet[2506]: W0414 13:32:11.601117 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.602003 kubelet[2506]: E0414 13:32:11.601396 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.616168 containerd[1454]: time="2026-04-14T13:32:11.615670285Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:32:11.616168 containerd[1454]: time="2026-04-14T13:32:11.615856372Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:32:11.616168 containerd[1454]: time="2026-04-14T13:32:11.615871632Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:32:11.616168 containerd[1454]: time="2026-04-14T13:32:11.616008826Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:32:11.689508 kubelet[2506]: E0414 13:32:11.689430 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:11.690170 kubelet[2506]: W0414 13:32:11.690017 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:11.690170 kubelet[2506]: E0414 13:32:11.690096 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:11.719339 systemd[1]: Started cri-containerd-109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc.scope - libcontainer container 109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc. Apr 14 13:32:11.779664 containerd[1454]: time="2026-04-14T13:32:11.779523120Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6847b996d8-mh4vw,Uid:30dd40e2-34af-4d02-ab75-c1f4d0654047,Namespace:calico-system,Attempt:0,} returns sandbox id \"9ecbdd7bf008eb04191e19cf678280ce120b63d7a524ede09b829a14283e529c\"" Apr 14 13:32:11.784908 kubelet[2506]: E0414 13:32:11.782350 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:32:11.788412 containerd[1454]: time="2026-04-14T13:32:11.786124038Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\"" Apr 14 13:32:11.915837 containerd[1454]: time="2026-04-14T13:32:11.915757630Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ch7dd,Uid:f458aee8-675f-406a-b973-ea55edc77b9d,Namespace:calico-system,Attempt:0,} returns sandbox id \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\"" Apr 14 13:32:13.482104 kubelet[2506]: E0414 13:32:13.482021 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:13.821523 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1832123473.mount: Deactivated successfully. Apr 14 13:32:15.496765 kubelet[2506]: E0414 13:32:15.496589 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:16.177188 containerd[1454]: time="2026-04-14T13:32:16.176187198Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:16.180418 containerd[1454]: time="2026-04-14T13:32:16.180367309Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.4: active requests=0, bytes read=36107596" Apr 14 13:32:16.183171 containerd[1454]: time="2026-04-14T13:32:16.183016064Z" level=info msg="ImageCreate event name:\"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:16.193562 containerd[1454]: time="2026-04-14T13:32:16.193444128Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:16.194545 containerd[1454]: time="2026-04-14T13:32:16.194196531Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.4\" with image id \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\", size \"36107450\" in 4.407641975s" Apr 14 13:32:16.195166 containerd[1454]: time="2026-04-14T13:32:16.194597268Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\" returns image reference \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\"" Apr 14 13:32:16.232763 containerd[1454]: time="2026-04-14T13:32:16.232659726Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\"" Apr 14 13:32:16.286835 containerd[1454]: time="2026-04-14T13:32:16.285937473Z" level=info msg="CreateContainer within sandbox \"9ecbdd7bf008eb04191e19cf678280ce120b63d7a524ede09b829a14283e529c\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Apr 14 13:32:16.327356 containerd[1454]: time="2026-04-14T13:32:16.326071986Z" level=info msg="CreateContainer within sandbox \"9ecbdd7bf008eb04191e19cf678280ce120b63d7a524ede09b829a14283e529c\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"73df5f8ba045c3864bf7ca89d1c2ae9b7c0fe2c0089a30f1e444c2e58694451d\"" Apr 14 13:32:16.333736 containerd[1454]: time="2026-04-14T13:32:16.333618311Z" level=info msg="StartContainer for \"73df5f8ba045c3864bf7ca89d1c2ae9b7c0fe2c0089a30f1e444c2e58694451d\"" Apr 14 13:32:16.489103 systemd[1]: Started cri-containerd-73df5f8ba045c3864bf7ca89d1c2ae9b7c0fe2c0089a30f1e444c2e58694451d.scope - libcontainer container 73df5f8ba045c3864bf7ca89d1c2ae9b7c0fe2c0089a30f1e444c2e58694451d. Apr 14 13:32:16.567149 containerd[1454]: time="2026-04-14T13:32:16.567083413Z" level=info msg="StartContainer for \"73df5f8ba045c3864bf7ca89d1c2ae9b7c0fe2c0089a30f1e444c2e58694451d\" returns successfully" Apr 14 13:32:17.027626 kubelet[2506]: E0414 13:32:17.027519 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:32:17.097140 kubelet[2506]: E0414 13:32:17.096719 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.097140 kubelet[2506]: W0414 13:32:17.096746 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.097140 kubelet[2506]: E0414 13:32:17.096810 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.099690 kubelet[2506]: E0414 13:32:17.097272 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.099690 kubelet[2506]: W0414 13:32:17.097282 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.099690 kubelet[2506]: E0414 13:32:17.097299 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.099690 kubelet[2506]: E0414 13:32:17.099166 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.099690 kubelet[2506]: W0414 13:32:17.099249 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.099690 kubelet[2506]: E0414 13:32:17.099345 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.099971 kubelet[2506]: E0414 13:32:17.099940 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.100027 kubelet[2506]: W0414 13:32:17.099961 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.100027 kubelet[2506]: E0414 13:32:17.100021 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.107464 kubelet[2506]: E0414 13:32:17.107175 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.110996 kubelet[2506]: W0414 13:32:17.107324 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.110996 kubelet[2506]: E0414 13:32:17.109900 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.112877 kubelet[2506]: E0414 13:32:17.112788 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.112877 kubelet[2506]: W0414 13:32:17.112870 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.112968 kubelet[2506]: E0414 13:32:17.112936 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.115279 kubelet[2506]: E0414 13:32:17.114308 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.116323 kubelet[2506]: W0414 13:32:17.115452 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.116323 kubelet[2506]: E0414 13:32:17.115524 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.123384 kubelet[2506]: E0414 13:32:17.123337 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.123561 kubelet[2506]: W0414 13:32:17.123551 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.125190 kubelet[2506]: E0414 13:32:17.124287 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.131736 kubelet[2506]: E0414 13:32:17.131681 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.134111 kubelet[2506]: W0414 13:32:17.133094 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.168352 kubelet[2506]: E0414 13:32:17.165203 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.177759 kubelet[2506]: E0414 13:32:17.176175 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.177759 kubelet[2506]: W0414 13:32:17.176291 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.177759 kubelet[2506]: E0414 13:32:17.176351 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.178171 kubelet[2506]: E0414 13:32:17.178079 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.178171 kubelet[2506]: W0414 13:32:17.178112 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.178171 kubelet[2506]: E0414 13:32:17.178160 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.178512 kubelet[2506]: I0414 13:32:17.178392 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-typha-6847b996d8-mh4vw" podStartSLOduration=2.730772859 podStartE2EDuration="7.178372041s" podCreationTimestamp="2026-04-14 13:32:10 +0000 UTC" firstStartedPulling="2026-04-14 13:32:11.78319478 +0000 UTC m=+21.529201161" lastFinishedPulling="2026-04-14 13:32:16.23079395 +0000 UTC m=+25.976800343" observedRunningTime="2026-04-14 13:32:17.155621 +0000 UTC m=+26.901627386" watchObservedRunningTime="2026-04-14 13:32:17.178372041 +0000 UTC m=+26.924378422" Apr 14 13:32:17.181187 kubelet[2506]: E0414 13:32:17.181060 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.181187 kubelet[2506]: W0414 13:32:17.181149 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.183312 kubelet[2506]: E0414 13:32:17.181212 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.183312 kubelet[2506]: E0414 13:32:17.182433 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.183312 kubelet[2506]: W0414 13:32:17.182441 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.183312 kubelet[2506]: E0414 13:32:17.182481 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.185419 kubelet[2506]: E0414 13:32:17.185369 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.185419 kubelet[2506]: W0414 13:32:17.185429 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.185772 kubelet[2506]: E0414 13:32:17.185469 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.188800 kubelet[2506]: E0414 13:32:17.185957 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.189803 kubelet[2506]: W0414 13:32:17.188999 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.221350 kubelet[2506]: E0414 13:32:17.219899 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.221350 kubelet[2506]: E0414 13:32:17.220499 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.221350 kubelet[2506]: W0414 13:32:17.220510 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.221350 kubelet[2506]: E0414 13:32:17.220559 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.222185 kubelet[2506]: E0414 13:32:17.222099 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.222214 kubelet[2506]: W0414 13:32:17.222183 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.222293 kubelet[2506]: E0414 13:32:17.222270 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.223050 kubelet[2506]: E0414 13:32:17.222982 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.223050 kubelet[2506]: W0414 13:32:17.223034 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.223248 kubelet[2506]: E0414 13:32:17.223073 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.223383 kubelet[2506]: E0414 13:32:17.223356 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.223383 kubelet[2506]: W0414 13:32:17.223379 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.223425 kubelet[2506]: E0414 13:32:17.223387 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.223560 kubelet[2506]: E0414 13:32:17.223531 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.223560 kubelet[2506]: W0414 13:32:17.223554 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.223604 kubelet[2506]: E0414 13:32:17.223564 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.223729 kubelet[2506]: E0414 13:32:17.223707 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.223751 kubelet[2506]: W0414 13:32:17.223728 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.223751 kubelet[2506]: E0414 13:32:17.223739 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.223894 kubelet[2506]: E0414 13:32:17.223869 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.223914 kubelet[2506]: W0414 13:32:17.223896 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.223979 kubelet[2506]: E0414 13:32:17.223903 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.226526 kubelet[2506]: E0414 13:32:17.225637 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.226526 kubelet[2506]: W0414 13:32:17.225677 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.226526 kubelet[2506]: E0414 13:32:17.225711 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.232478 kubelet[2506]: E0414 13:32:17.228925 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.232478 kubelet[2506]: W0414 13:32:17.229134 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.232478 kubelet[2506]: E0414 13:32:17.229185 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.234958 kubelet[2506]: E0414 13:32:17.234327 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.234958 kubelet[2506]: W0414 13:32:17.234370 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.234958 kubelet[2506]: E0414 13:32:17.234462 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.235305 kubelet[2506]: E0414 13:32:17.235276 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.235378 kubelet[2506]: W0414 13:32:17.235339 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.235480 kubelet[2506]: E0414 13:32:17.235469 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.238061 kubelet[2506]: E0414 13:32:17.237998 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.238061 kubelet[2506]: W0414 13:32:17.238040 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.238061 kubelet[2506]: E0414 13:32:17.238065 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.238917 kubelet[2506]: E0414 13:32:17.238887 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.238917 kubelet[2506]: W0414 13:32:17.238905 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.238917 kubelet[2506]: E0414 13:32:17.238916 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.239139 kubelet[2506]: E0414 13:32:17.239105 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.239139 kubelet[2506]: W0414 13:32:17.239135 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.239177 kubelet[2506]: E0414 13:32:17.239146 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.239454 kubelet[2506]: E0414 13:32:17.239367 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.239454 kubelet[2506]: W0414 13:32:17.239441 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.239454 kubelet[2506]: E0414 13:32:17.239456 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.241366 kubelet[2506]: E0414 13:32:17.239714 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.241366 kubelet[2506]: W0414 13:32:17.239732 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.241366 kubelet[2506]: E0414 13:32:17.239740 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.241366 kubelet[2506]: E0414 13:32:17.240265 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.241366 kubelet[2506]: W0414 13:32:17.240272 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.241366 kubelet[2506]: E0414 13:32:17.240280 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.249937 kubelet[2506]: E0414 13:32:17.249753 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:17.249937 kubelet[2506]: W0414 13:32:17.249931 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:17.250131 kubelet[2506]: E0414 13:32:17.249998 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:17.484403 kubelet[2506]: E0414 13:32:17.484304 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:18.075447 kubelet[2506]: E0414 13:32:18.074114 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:32:18.101086 kubelet[2506]: E0414 13:32:18.100819 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.101086 kubelet[2506]: W0414 13:32:18.100899 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.101086 kubelet[2506]: E0414 13:32:18.100960 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.101575 kubelet[2506]: E0414 13:32:18.101311 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.101575 kubelet[2506]: W0414 13:32:18.101321 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.101575 kubelet[2506]: E0414 13:32:18.101334 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.152122 kubelet[2506]: E0414 13:32:18.151439 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.152122 kubelet[2506]: W0414 13:32:18.151567 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.152122 kubelet[2506]: E0414 13:32:18.151734 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.153006 kubelet[2506]: E0414 13:32:18.152864 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.153006 kubelet[2506]: W0414 13:32:18.152883 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.153006 kubelet[2506]: E0414 13:32:18.152901 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.156857 kubelet[2506]: E0414 13:32:18.154920 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.156857 kubelet[2506]: W0414 13:32:18.155041 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.156857 kubelet[2506]: E0414 13:32:18.155808 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.157960 kubelet[2506]: E0414 13:32:18.157844 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.158200 kubelet[2506]: W0414 13:32:18.158147 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.158371 kubelet[2506]: E0414 13:32:18.158350 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.160289 kubelet[2506]: E0414 13:32:18.159698 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.161466 kubelet[2506]: W0414 13:32:18.161206 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.161766 kubelet[2506]: E0414 13:32:18.161650 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.172731 kubelet[2506]: E0414 13:32:18.172494 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.173456 kubelet[2506]: W0414 13:32:18.173276 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.173658 kubelet[2506]: E0414 13:32:18.173611 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.174437 kubelet[2506]: E0414 13:32:18.174422 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.174506 kubelet[2506]: W0414 13:32:18.174496 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.174553 kubelet[2506]: E0414 13:32:18.174545 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.178870 kubelet[2506]: E0414 13:32:18.178756 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.183128 kubelet[2506]: W0414 13:32:18.181474 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.183128 kubelet[2506]: E0414 13:32:18.181884 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.190316 kubelet[2506]: E0414 13:32:18.187766 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.191864 kubelet[2506]: W0414 13:32:18.187898 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.194560 kubelet[2506]: E0414 13:32:18.193931 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.220764 kubelet[2506]: E0414 13:32:18.219845 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.222504 kubelet[2506]: W0414 13:32:18.221913 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.222504 kubelet[2506]: E0414 13:32:18.222296 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.229356 kubelet[2506]: E0414 13:32:18.226621 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.229356 kubelet[2506]: W0414 13:32:18.227781 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.235672 kubelet[2506]: E0414 13:32:18.233897 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.244136 kubelet[2506]: E0414 13:32:18.243838 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.246871 kubelet[2506]: W0414 13:32:18.246332 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.250873 kubelet[2506]: E0414 13:32:18.250636 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.278795 kubelet[2506]: E0414 13:32:18.276915 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.282074 kubelet[2506]: W0414 13:32:18.281859 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.282413 kubelet[2506]: E0414 13:32:18.282174 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.292396 kubelet[2506]: E0414 13:32:18.290901 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.292396 kubelet[2506]: W0414 13:32:18.291744 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.292396 kubelet[2506]: E0414 13:32:18.292041 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.296472 kubelet[2506]: E0414 13:32:18.295985 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.296987 kubelet[2506]: W0414 13:32:18.296865 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.355974 kubelet[2506]: E0414 13:32:18.354733 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.364962 kubelet[2506]: E0414 13:32:18.363866 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.369305 kubelet[2506]: W0414 13:32:18.366856 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.369633 kubelet[2506]: E0414 13:32:18.369170 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.371484 kubelet[2506]: E0414 13:32:18.371421 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.375035 kubelet[2506]: W0414 13:32:18.372570 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.376511 kubelet[2506]: E0414 13:32:18.376106 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.376788 kubelet[2506]: E0414 13:32:18.376777 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.376875 kubelet[2506]: W0414 13:32:18.376832 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.376875 kubelet[2506]: E0414 13:32:18.376846 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.377431 kubelet[2506]: E0414 13:32:18.377393 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.377431 kubelet[2506]: W0414 13:32:18.377421 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.377483 kubelet[2506]: E0414 13:32:18.377436 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.386154 kubelet[2506]: E0414 13:32:18.385881 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.386739 kubelet[2506]: W0414 13:32:18.386570 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.386979 kubelet[2506]: E0414 13:32:18.386800 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.393324 kubelet[2506]: E0414 13:32:18.392790 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.393324 kubelet[2506]: W0414 13:32:18.393209 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.393597 kubelet[2506]: E0414 13:32:18.393402 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.394311 kubelet[2506]: E0414 13:32:18.393911 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.394311 kubelet[2506]: W0414 13:32:18.393921 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.394311 kubelet[2506]: E0414 13:32:18.393931 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.394311 kubelet[2506]: E0414 13:32:18.394214 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.394311 kubelet[2506]: W0414 13:32:18.394309 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.394483 kubelet[2506]: E0414 13:32:18.394319 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.400151 kubelet[2506]: E0414 13:32:18.400043 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.400151 kubelet[2506]: W0414 13:32:18.400116 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.400443 kubelet[2506]: E0414 13:32:18.400170 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.425180 kubelet[2506]: E0414 13:32:18.401779 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.426992 kubelet[2506]: W0414 13:32:18.426791 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.427208 kubelet[2506]: E0414 13:32:18.427126 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.427770 kubelet[2506]: E0414 13:32:18.427694 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.427770 kubelet[2506]: W0414 13:32:18.427748 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.427770 kubelet[2506]: E0414 13:32:18.427762 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.432062 kubelet[2506]: E0414 13:32:18.431775 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.434856 kubelet[2506]: W0414 13:32:18.432773 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.434856 kubelet[2506]: E0414 13:32:18.434372 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.439447 kubelet[2506]: E0414 13:32:18.438986 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.440255 kubelet[2506]: W0414 13:32:18.439813 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.444122 kubelet[2506]: E0414 13:32:18.444038 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.462004 kubelet[2506]: E0414 13:32:18.461709 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.462574 kubelet[2506]: W0414 13:32:18.462056 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.463831 kubelet[2506]: E0414 13:32:18.463549 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.468798 kubelet[2506]: E0414 13:32:18.468561 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.468798 kubelet[2506]: W0414 13:32:18.468645 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.468798 kubelet[2506]: E0414 13:32:18.468743 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.476166 kubelet[2506]: E0414 13:32:18.476065 2506 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 14 13:32:18.476668 kubelet[2506]: W0414 13:32:18.476175 2506 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 14 13:32:18.477303 kubelet[2506]: E0414 13:32:18.477074 2506 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 14 13:32:18.772295 containerd[1454]: time="2026-04-14T13:32:18.769899745Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:18.772295 containerd[1454]: time="2026-04-14T13:32:18.772074751Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4: active requests=0, bytes read=4630250" Apr 14 13:32:18.773592 containerd[1454]: time="2026-04-14T13:32:18.773527043Z" level=info msg="ImageCreate event name:\"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:18.778087 containerd[1454]: time="2026-04-14T13:32:18.777985770Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:18.779470 containerd[1454]: time="2026-04-14T13:32:18.779404861Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" with image id \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\", size \"6186255\" in 2.546677082s" Apr 14 13:32:18.779470 containerd[1454]: time="2026-04-14T13:32:18.779464340Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" returns image reference \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\"" Apr 14 13:32:18.802824 containerd[1454]: time="2026-04-14T13:32:18.802705316Z" level=info msg="CreateContainer within sandbox \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Apr 14 13:32:18.838110 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount467136188.mount: Deactivated successfully. Apr 14 13:32:18.845889 containerd[1454]: time="2026-04-14T13:32:18.845797828Z" level=info msg="CreateContainer within sandbox \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c\"" Apr 14 13:32:18.848166 containerd[1454]: time="2026-04-14T13:32:18.848062825Z" level=info msg="StartContainer for \"9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c\"" Apr 14 13:32:18.972735 systemd[1]: Started cri-containerd-9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c.scope - libcontainer container 9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c. Apr 14 13:32:19.050567 containerd[1454]: time="2026-04-14T13:32:19.050391617Z" level=info msg="StartContainer for \"9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c\" returns successfully" Apr 14 13:32:19.064021 systemd[1]: cri-containerd-9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c.scope: Deactivated successfully. Apr 14 13:32:19.124448 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c-rootfs.mount: Deactivated successfully. Apr 14 13:32:19.135033 containerd[1454]: time="2026-04-14T13:32:19.134924435Z" level=info msg="shim disconnected" id=9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c namespace=k8s.io Apr 14 13:32:19.135033 containerd[1454]: time="2026-04-14T13:32:19.135003373Z" level=warning msg="cleaning up after shim disconnected" id=9f90af7d0ea52be3b972884eaebf3439e1bc8aa76bd830ed0c073b9051c6117c namespace=k8s.io Apr 14 13:32:19.135033 containerd[1454]: time="2026-04-14T13:32:19.135014351Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 14 13:32:19.511001 kubelet[2506]: E0414 13:32:19.501412 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:20.072102 containerd[1454]: time="2026-04-14T13:32:20.071958290Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\"" Apr 14 13:32:21.482608 kubelet[2506]: E0414 13:32:21.482461 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:23.494861 kubelet[2506]: E0414 13:32:23.483012 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:25.496789 kubelet[2506]: E0414 13:32:25.496208 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:27.483841 kubelet[2506]: E0414 13:32:27.483469 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:29.502024 kubelet[2506]: E0414 13:32:29.501932 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:31.555473 kubelet[2506]: E0414 13:32:31.555140 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:33.485330 kubelet[2506]: E0414 13:32:33.484466 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:35.487144 kubelet[2506]: E0414 13:32:35.487029 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:35.946928 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1611838290.mount: Deactivated successfully. Apr 14 13:32:36.142425 containerd[1454]: time="2026-04-14T13:32:36.141833359Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:36.146125 containerd[1454]: time="2026-04-14T13:32:36.145673693Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.4: active requests=0, bytes read=159838564" Apr 14 13:32:36.150765 containerd[1454]: time="2026-04-14T13:32:36.150683320Z" level=info msg="ImageCreate event name:\"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:36.161754 containerd[1454]: time="2026-04-14T13:32:36.159100198Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:36.167469 containerd[1454]: time="2026-04-14T13:32:36.167408848Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.4\" with image id \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\", size \"159838426\" in 16.095321542s" Apr 14 13:32:36.167469 containerd[1454]: time="2026-04-14T13:32:36.167463328Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\" returns image reference \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\"" Apr 14 13:32:36.214351 containerd[1454]: time="2026-04-14T13:32:36.213460642Z" level=info msg="CreateContainer within sandbox \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Apr 14 13:32:36.348132 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2891921899.mount: Deactivated successfully. Apr 14 13:32:36.399434 containerd[1454]: time="2026-04-14T13:32:36.399161177Z" level=info msg="CreateContainer within sandbox \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687\"" Apr 14 13:32:36.404768 containerd[1454]: time="2026-04-14T13:32:36.402735407Z" level=info msg="StartContainer for \"ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687\"" Apr 14 13:32:36.763421 systemd[1]: Started cri-containerd-ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687.scope - libcontainer container ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687. Apr 14 13:32:37.066157 containerd[1454]: time="2026-04-14T13:32:37.065575440Z" level=info msg="StartContainer for \"ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687\" returns successfully" Apr 14 13:32:37.397685 systemd[1]: cri-containerd-ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687.scope: Deactivated successfully. Apr 14 13:32:37.487526 kubelet[2506]: E0414 13:32:37.487386 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:37.574955 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687-rootfs.mount: Deactivated successfully. Apr 14 13:32:37.588887 containerd[1454]: time="2026-04-14T13:32:37.588584374Z" level=info msg="shim disconnected" id=ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687 namespace=k8s.io Apr 14 13:32:37.588887 containerd[1454]: time="2026-04-14T13:32:37.588869212Z" level=warning msg="cleaning up after shim disconnected" id=ae1a59fa383998b803c3f8bffce4aec052a528865bc0da2557ff423ea31cd687 namespace=k8s.io Apr 14 13:32:37.588887 containerd[1454]: time="2026-04-14T13:32:37.588901693Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 14 13:32:38.508178 containerd[1454]: time="2026-04-14T13:32:38.508123304Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\"" Apr 14 13:32:39.489158 kubelet[2506]: E0414 13:32:39.488998 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:41.482652 kubelet[2506]: E0414 13:32:41.482562 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:43.483810 kubelet[2506]: E0414 13:32:43.483575 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:44.846391 containerd[1454]: time="2026-04-14T13:32:44.845936279Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:44.848503 containerd[1454]: time="2026-04-14T13:32:44.847974198Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.4: active requests=0, bytes read=70611671" Apr 14 13:32:44.849755 containerd[1454]: time="2026-04-14T13:32:44.849698088Z" level=info msg="ImageCreate event name:\"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:44.856012 containerd[1454]: time="2026-04-14T13:32:44.855930213Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:32:44.857913 containerd[1454]: time="2026-04-14T13:32:44.857383628Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.4\" with image id \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\", size \"72167716\" in 6.349223597s" Apr 14 13:32:44.857913 containerd[1454]: time="2026-04-14T13:32:44.857423422Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\" returns image reference \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\"" Apr 14 13:32:44.957895 containerd[1454]: time="2026-04-14T13:32:44.957819158Z" level=info msg="CreateContainer within sandbox \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Apr 14 13:32:45.045330 containerd[1454]: time="2026-04-14T13:32:45.045211941Z" level=info msg="CreateContainer within sandbox \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58\"" Apr 14 13:32:45.055523 containerd[1454]: time="2026-04-14T13:32:45.055374047Z" level=info msg="StartContainer for \"1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58\"" Apr 14 13:32:45.242371 systemd[1]: Started cri-containerd-1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58.scope - libcontainer container 1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58. Apr 14 13:32:45.384817 containerd[1454]: time="2026-04-14T13:32:45.384754129Z" level=info msg="StartContainer for \"1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58\" returns successfully" Apr 14 13:32:45.486035 kubelet[2506]: E0414 13:32:45.485840 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:47.489269 kubelet[2506]: E0414 13:32:47.484150 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:47.618576 systemd[1]: cri-containerd-1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58.scope: Deactivated successfully. Apr 14 13:32:47.619850 systemd[1]: cri-containerd-1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58.scope: Consumed 1.587s CPU time. Apr 14 13:32:47.795076 kubelet[2506]: I0414 13:32:47.794646 2506 kubelet_node_status.go:427] "Fast updating node status as it just became ready" Apr 14 13:32:47.835530 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58-rootfs.mount: Deactivated successfully. Apr 14 13:32:47.887372 containerd[1454]: time="2026-04-14T13:32:47.886628172Z" level=info msg="shim disconnected" id=1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58 namespace=k8s.io Apr 14 13:32:47.887372 containerd[1454]: time="2026-04-14T13:32:47.886735371Z" level=warning msg="cleaning up after shim disconnected" id=1be4b23325b5de520ab9fa0e6aa167620fb9dbdd48f8503f7b9a2db6aeee9e58 namespace=k8s.io Apr 14 13:32:47.887372 containerd[1454]: time="2026-04-14T13:32:47.886746216Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 14 13:32:48.071416 kubelet[2506]: E0414 13:32:48.070831 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:32:48.507403 systemd[1]: Created slice kubepods-burstable-pod9395319e_7ca7_4248_8723_a075a8a69da5.slice - libcontainer container kubepods-burstable-pod9395319e_7ca7_4248_8723_a075a8a69da5.slice. Apr 14 13:32:48.569708 kubelet[2506]: I0414 13:32:48.569177 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9395319e-7ca7-4248-8723-a075a8a69da5-config-volume\") pod \"coredns-7d764666f9-km2ks\" (UID: \"9395319e-7ca7-4248-8723-a075a8a69da5\") " pod="kube-system/coredns-7d764666f9-km2ks" Apr 14 13:32:48.574851 kubelet[2506]: I0414 13:32:48.574523 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv94m\" (UniqueName: \"kubernetes.io/projected/9395319e-7ca7-4248-8723-a075a8a69da5-kube-api-access-hv94m\") pod \"coredns-7d764666f9-km2ks\" (UID: \"9395319e-7ca7-4248-8723-a075a8a69da5\") " pod="kube-system/coredns-7d764666f9-km2ks" Apr 14 13:32:48.579464 kubelet[2506]: E0414 13:32:48.574627 2506 status_manager.go:1045] "Failed to get status for pod" err="pods \"coredns-7d764666f9-km2ks\" is forbidden: User \"system:node:localhost\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'localhost' and this object" podUID="9395319e-7ca7-4248-8723-a075a8a69da5" pod="kube-system/coredns-7d764666f9-km2ks" Apr 14 13:32:48.776836 systemd[1]: Created slice kubepods-besteffort-pod2bca4db2_bc4e_488c_88eb_36317594e234.slice - libcontainer container kubepods-besteffort-pod2bca4db2_bc4e_488c_88eb_36317594e234.slice. Apr 14 13:32:48.888577 kubelet[2506]: I0414 13:32:48.807795 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2bca4db2-bc4e-488c-88eb-36317594e234-tigera-ca-bundle\") pod \"calico-kube-controllers-5bb46d7bb5-xwkl7\" (UID: \"2bca4db2-bc4e-488c-88eb-36317594e234\") " pod="calico-system/calico-kube-controllers-5bb46d7bb5-xwkl7" Apr 14 13:32:48.897484 kubelet[2506]: I0414 13:32:48.897139 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4nz4\" (UniqueName: \"kubernetes.io/projected/2bca4db2-bc4e-488c-88eb-36317594e234-kube-api-access-s4nz4\") pod \"calico-kube-controllers-5bb46d7bb5-xwkl7\" (UID: \"2bca4db2-bc4e-488c-88eb-36317594e234\") " pod="calico-system/calico-kube-controllers-5bb46d7bb5-xwkl7" Apr 14 13:32:48.991787 systemd[1]: Created slice kubepods-burstable-pod8d625da5_d2c2_45ba_9200_2567610f3552.slice - libcontainer container kubepods-burstable-pod8d625da5_d2c2_45ba_9200_2567610f3552.slice. Apr 14 13:32:49.065599 systemd[1]: Created slice kubepods-besteffort-pod10f4fd58_e76c_4b5c_a2bd_c998eae079c3.slice - libcontainer container kubepods-besteffort-pod10f4fd58_e76c_4b5c_a2bd_c998eae079c3.slice. Apr 14 13:32:49.096340 kubelet[2506]: I0414 13:32:49.095606 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-ca-bundle\") pod \"whisker-55d9ccd5cb-vc6bk\" (UID: \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\") " pod="calico-system/whisker-55d9ccd5cb-vc6bk" Apr 14 13:32:49.096340 kubelet[2506]: I0414 13:32:49.095702 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm46f\" (UniqueName: \"kubernetes.io/projected/8d625da5-d2c2-45ba-9200-2567610f3552-kube-api-access-cm46f\") pod \"coredns-7d764666f9-m5vbx\" (UID: \"8d625da5-d2c2-45ba-9200-2567610f3552\") " pod="kube-system/coredns-7d764666f9-m5vbx" Apr 14 13:32:49.096340 kubelet[2506]: I0414 13:32:49.095720 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-backend-key-pair\") pod \"whisker-55d9ccd5cb-vc6bk\" (UID: \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\") " pod="calico-system/whisker-55d9ccd5cb-vc6bk" Apr 14 13:32:49.096340 kubelet[2506]: I0414 13:32:49.095761 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d625da5-d2c2-45ba-9200-2567610f3552-config-volume\") pod \"coredns-7d764666f9-m5vbx\" (UID: \"8d625da5-d2c2-45ba-9200-2567610f3552\") " pod="kube-system/coredns-7d764666f9-m5vbx" Apr 14 13:32:49.096340 kubelet[2506]: I0414 13:32:49.095776 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4b5j\" (UniqueName: \"kubernetes.io/projected/8a9902ab-3b55-411b-9c72-16f6b24f69c0-kube-api-access-s4b5j\") pod \"whisker-55d9ccd5cb-vc6bk\" (UID: \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\") " pod="calico-system/whisker-55d9ccd5cb-vc6bk" Apr 14 13:32:49.096946 kubelet[2506]: I0414 13:32:49.096404 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-nginx-config\") pod \"whisker-55d9ccd5cb-vc6bk\" (UID: \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\") " pod="calico-system/whisker-55d9ccd5cb-vc6bk" Apr 14 13:32:49.203816 systemd[1]: Created slice kubepods-besteffort-pod8a9902ab_3b55_411b_9c72_16f6b24f69c0.slice - libcontainer container kubepods-besteffort-pod8a9902ab_3b55_411b_9c72_16f6b24f69c0.slice. Apr 14 13:32:49.207593 kubelet[2506]: I0414 13:32:49.205102 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10f4fd58-e76c-4b5c-a2bd-c998eae079c3-goldmane-ca-bundle\") pod \"goldmane-9f7667bb8-268pt\" (UID: \"10f4fd58-e76c-4b5c-a2bd-c998eae079c3\") " pod="calico-system/goldmane-9f7667bb8-268pt" Apr 14 13:32:49.207593 kubelet[2506]: I0414 13:32:49.205178 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwmlm\" (UniqueName: \"kubernetes.io/projected/10f4fd58-e76c-4b5c-a2bd-c998eae079c3-kube-api-access-lwmlm\") pod \"goldmane-9f7667bb8-268pt\" (UID: \"10f4fd58-e76c-4b5c-a2bd-c998eae079c3\") " pod="calico-system/goldmane-9f7667bb8-268pt" Apr 14 13:32:49.207593 kubelet[2506]: I0414 13:32:49.205194 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqjdb\" (UniqueName: \"kubernetes.io/projected/ada6aff6-b823-4c1b-8ba1-78289ce2a978-kube-api-access-gqjdb\") pod \"calico-apiserver-56fd9bc5cd-hqzvr\" (UID: \"ada6aff6-b823-4c1b-8ba1-78289ce2a978\") " pod="calico-system/calico-apiserver-56fd9bc5cd-hqzvr" Apr 14 13:32:49.232064 kubelet[2506]: E0414 13:32:49.231986 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:32:49.239119 kubelet[2506]: I0414 13:32:49.238851 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/10f4fd58-e76c-4b5c-a2bd-c998eae079c3-goldmane-key-pair\") pod \"goldmane-9f7667bb8-268pt\" (UID: \"10f4fd58-e76c-4b5c-a2bd-c998eae079c3\") " pod="calico-system/goldmane-9f7667bb8-268pt" Apr 14 13:32:49.239286 kubelet[2506]: I0414 13:32:49.239255 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10f4fd58-e76c-4b5c-a2bd-c998eae079c3-config\") pod \"goldmane-9f7667bb8-268pt\" (UID: \"10f4fd58-e76c-4b5c-a2bd-c998eae079c3\") " pod="calico-system/goldmane-9f7667bb8-268pt" Apr 14 13:32:49.239312 kubelet[2506]: I0414 13:32:49.239285 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f6f60262-e827-44cc-9078-cdb0fd928b6c-calico-apiserver-certs\") pod \"calico-apiserver-56fd9bc5cd-62gjx\" (UID: \"f6f60262-e827-44cc-9078-cdb0fd928b6c\") " pod="calico-system/calico-apiserver-56fd9bc5cd-62gjx" Apr 14 13:32:49.239337 kubelet[2506]: I0414 13:32:49.239328 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/ada6aff6-b823-4c1b-8ba1-78289ce2a978-calico-apiserver-certs\") pod \"calico-apiserver-56fd9bc5cd-hqzvr\" (UID: \"ada6aff6-b823-4c1b-8ba1-78289ce2a978\") " pod="calico-system/calico-apiserver-56fd9bc5cd-hqzvr" Apr 14 13:32:49.239417 kubelet[2506]: I0414 13:32:49.239400 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8ftn\" (UniqueName: \"kubernetes.io/projected/f6f60262-e827-44cc-9078-cdb0fd928b6c-kube-api-access-g8ftn\") pod \"calico-apiserver-56fd9bc5cd-62gjx\" (UID: \"f6f60262-e827-44cc-9078-cdb0fd928b6c\") " pod="calico-system/calico-apiserver-56fd9bc5cd-62gjx" Apr 14 13:32:49.241484 containerd[1454]: time="2026-04-14T13:32:49.239745934Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-km2ks,Uid:9395319e-7ca7-4248-8723-a075a8a69da5,Namespace:kube-system,Attempt:0,}" Apr 14 13:32:49.262157 systemd[1]: Created slice kubepods-besteffort-podf6f60262_e827_44cc_9078_cdb0fd928b6c.slice - libcontainer container kubepods-besteffort-podf6f60262_e827_44cc_9078_cdb0fd928b6c.slice. Apr 14 13:32:49.271372 containerd[1454]: time="2026-04-14T13:32:49.269256278Z" level=info msg="CreateContainer within sandbox \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Apr 14 13:32:49.343537 systemd[1]: Created slice kubepods-besteffort-podada6aff6_b823_4c1b_8ba1_78289ce2a978.slice - libcontainer container kubepods-besteffort-podada6aff6_b823_4c1b_8ba1_78289ce2a978.slice. Apr 14 13:32:49.605457 containerd[1454]: time="2026-04-14T13:32:49.604318522Z" level=info msg="CreateContainer within sandbox \"109bda0a163cddf4f590f9777d9bd8aa752485f563924cca0d3325edc0dbbccc\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"bdc6f593c419f26c1f7641ebeaa5a32319685d6d23d132caf7c51c2491f9ae4c\"" Apr 14 13:32:49.617065 containerd[1454]: time="2026-04-14T13:32:49.616939108Z" level=info msg="StartContainer for \"bdc6f593c419f26c1f7641ebeaa5a32319685d6d23d132caf7c51c2491f9ae4c\"" Apr 14 13:32:49.700574 systemd[1]: Created slice kubepods-besteffort-podb5b0d33b_1b54_43cc_a4d3_cd788e8eb914.slice - libcontainer container kubepods-besteffort-podb5b0d33b_1b54_43cc_a4d3_cd788e8eb914.slice. Apr 14 13:32:49.856651 containerd[1454]: time="2026-04-14T13:32:49.855564565Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5xvdj,Uid:b5b0d33b-1b54-43cc-a4d3-cd788e8eb914,Namespace:calico-system,Attempt:0,}" Apr 14 13:32:49.888640 systemd[1]: Started cri-containerd-bdc6f593c419f26c1f7641ebeaa5a32319685d6d23d132caf7c51c2491f9ae4c.scope - libcontainer container bdc6f593c419f26c1f7641ebeaa5a32319685d6d23d132caf7c51c2491f9ae4c. Apr 14 13:32:49.915324 containerd[1454]: time="2026-04-14T13:32:49.911608758Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bb46d7bb5-xwkl7,Uid:2bca4db2-bc4e-488c-88eb-36317594e234,Namespace:calico-system,Attempt:0,}" Apr 14 13:32:49.919527 containerd[1454]: time="2026-04-14T13:32:49.919471228Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56fd9bc5cd-62gjx,Uid:f6f60262-e827-44cc-9078-cdb0fd928b6c,Namespace:calico-system,Attempt:0,}" Apr 14 13:32:49.920260 containerd[1454]: time="2026-04-14T13:32:49.919710103Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-55d9ccd5cb-vc6bk,Uid:8a9902ab-3b55-411b-9c72-16f6b24f69c0,Namespace:calico-system,Attempt:0,}" Apr 14 13:32:49.936629 kubelet[2506]: E0414 13:32:49.934478 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:32:49.954658 containerd[1454]: time="2026-04-14T13:32:49.954274180Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-m5vbx,Uid:8d625da5-d2c2-45ba-9200-2567610f3552,Namespace:kube-system,Attempt:0,}" Apr 14 13:32:50.057280 containerd[1454]: time="2026-04-14T13:32:50.057185769Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56fd9bc5cd-hqzvr,Uid:ada6aff6-b823-4c1b-8ba1-78289ce2a978,Namespace:calico-system,Attempt:0,}" Apr 14 13:32:50.187333 containerd[1454]: time="2026-04-14T13:32:50.177511759Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-9f7667bb8-268pt,Uid:10f4fd58-e76c-4b5c-a2bd-c998eae079c3,Namespace:calico-system,Attempt:0,}" Apr 14 13:32:50.705581 containerd[1454]: time="2026-04-14T13:32:50.705497392Z" level=info msg="StartContainer for \"bdc6f593c419f26c1f7641ebeaa5a32319685d6d23d132caf7c51c2491f9ae4c\" returns successfully" Apr 14 13:32:50.756623 containerd[1454]: time="2026-04-14T13:32:50.740588523Z" level=error msg="Failed to destroy network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:50.756623 containerd[1454]: time="2026-04-14T13:32:50.743363549Z" level=error msg="encountered an error cleaning up failed sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:50.756623 containerd[1454]: time="2026-04-14T13:32:50.743482947Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-km2ks,Uid:9395319e-7ca7-4248-8723-a075a8a69da5,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:50.755993 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c-shm.mount: Deactivated successfully. Apr 14 13:32:50.767837 kubelet[2506]: E0414 13:32:50.767698 2506 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:50.767837 kubelet[2506]: E0414 13:32:50.767845 2506 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-km2ks" Apr 14 13:32:50.772982 kubelet[2506]: E0414 13:32:50.767867 2506 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-km2ks" Apr 14 13:32:50.772982 kubelet[2506]: E0414 13:32:50.772748 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7d764666f9-km2ks_kube-system(9395319e-7ca7-4248-8723-a075a8a69da5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7d764666f9-km2ks_kube-system(9395319e-7ca7-4248-8723-a075a8a69da5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7d764666f9-km2ks" podUID="9395319e-7ca7-4248-8723-a075a8a69da5" Apr 14 13:32:51.237695 containerd[1454]: time="2026-04-14T13:32:51.237134202Z" level=error msg="Failed to destroy network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.241687 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97-shm.mount: Deactivated successfully. Apr 14 13:32:51.254505 containerd[1454]: time="2026-04-14T13:32:51.253602960Z" level=error msg="encountered an error cleaning up failed sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.254505 containerd[1454]: time="2026-04-14T13:32:51.253877149Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5xvdj,Uid:b5b0d33b-1b54-43cc-a4d3-cd788e8eb914,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.255253 kubelet[2506]: E0414 13:32:51.254856 2506 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.255931 kubelet[2506]: E0414 13:32:51.255805 2506 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5xvdj" Apr 14 13:32:51.255931 kubelet[2506]: E0414 13:32:51.255896 2506 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5xvdj" Apr 14 13:32:51.256053 kubelet[2506]: E0414 13:32:51.255956 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-5xvdj_calico-system(b5b0d33b-1b54-43cc-a4d3-cd788e8eb914)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-5xvdj_calico-system(b5b0d33b-1b54-43cc-a4d3-cd788e8eb914)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:51.483616 containerd[1454]: time="2026-04-14T13:32:51.483041333Z" level=error msg="Failed to destroy network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.486869 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8-shm.mount: Deactivated successfully. Apr 14 13:32:51.587200 containerd[1454]: time="2026-04-14T13:32:51.585580738Z" level=error msg="encountered an error cleaning up failed sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.587200 containerd[1454]: time="2026-04-14T13:32:51.585734763Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56fd9bc5cd-62gjx,Uid:f6f60262-e827-44cc-9078-cdb0fd928b6c,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.613477 kubelet[2506]: E0414 13:32:51.589174 2506 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.632578 kubelet[2506]: E0414 13:32:51.612178 2506 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-56fd9bc5cd-62gjx" Apr 14 13:32:51.639293 kubelet[2506]: E0414 13:32:51.632721 2506 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-56fd9bc5cd-62gjx" Apr 14 13:32:51.639775 kubelet[2506]: E0414 13:32:51.639512 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-56fd9bc5cd-62gjx_calico-system(f6f60262-e827-44cc-9078-cdb0fd928b6c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-56fd9bc5cd-62gjx_calico-system(f6f60262-e827-44cc-9078-cdb0fd928b6c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-56fd9bc5cd-62gjx" podUID="f6f60262-e827-44cc-9078-cdb0fd928b6c" Apr 14 13:32:51.662830 containerd[1454]: time="2026-04-14T13:32:51.662723075Z" level=error msg="Failed to destroy network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.678791 containerd[1454]: time="2026-04-14T13:32:51.675984812Z" level=error msg="encountered an error cleaning up failed sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.681435 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9-shm.mount: Deactivated successfully. Apr 14 13:32:51.704150 containerd[1454]: time="2026-04-14T13:32:51.703767700Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-m5vbx,Uid:8d625da5-d2c2-45ba-9200-2567610f3552,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.792723 containerd[1454]: time="2026-04-14T13:32:51.702930810Z" level=error msg="Failed to destroy network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.793764 kubelet[2506]: E0414 13:32:51.786795 2506 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.793764 kubelet[2506]: E0414 13:32:51.789104 2506 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-m5vbx" Apr 14 13:32:51.793764 kubelet[2506]: E0414 13:32:51.790967 2506 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-m5vbx" Apr 14 13:32:51.794006 kubelet[2506]: E0414 13:32:51.791710 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7d764666f9-m5vbx_kube-system(8d625da5-d2c2-45ba-9200-2567610f3552)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7d764666f9-m5vbx_kube-system(8d625da5-d2c2-45ba-9200-2567610f3552)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7d764666f9-m5vbx" podUID="8d625da5-d2c2-45ba-9200-2567610f3552" Apr 14 13:32:51.796635 containerd[1454]: time="2026-04-14T13:32:51.795402057Z" level=error msg="encountered an error cleaning up failed sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.796635 containerd[1454]: time="2026-04-14T13:32:51.795557329Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bb46d7bb5-xwkl7,Uid:2bca4db2-bc4e-488c-88eb-36317594e234,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.797815 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970-shm.mount: Deactivated successfully. Apr 14 13:32:51.798020 kubelet[2506]: E0414 13:32:51.797609 2506 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.798020 kubelet[2506]: E0414 13:32:51.797880 2506 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bb46d7bb5-xwkl7" Apr 14 13:32:51.798020 kubelet[2506]: E0414 13:32:51.797906 2506 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bb46d7bb5-xwkl7" Apr 14 13:32:51.798186 kubelet[2506]: E0414 13:32:51.798045 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5bb46d7bb5-xwkl7_calico-system(2bca4db2-bc4e-488c-88eb-36317594e234)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5bb46d7bb5-xwkl7_calico-system(2bca4db2-bc4e-488c-88eb-36317594e234)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bb46d7bb5-xwkl7" podUID="2bca4db2-bc4e-488c-88eb-36317594e234" Apr 14 13:32:51.802071 containerd[1454]: time="2026-04-14T13:32:51.801563522Z" level=error msg="Failed to destroy network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.808858 containerd[1454]: time="2026-04-14T13:32:51.805131458Z" level=error msg="encountered an error cleaning up failed sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.808858 containerd[1454]: time="2026-04-14T13:32:51.808676659Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-55d9ccd5cb-vc6bk,Uid:8a9902ab-3b55-411b-9c72-16f6b24f69c0,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.825520 kubelet[2506]: E0414 13:32:51.819463 2506 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.848992 kubelet[2506]: E0414 13:32:51.830179 2506 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-55d9ccd5cb-vc6bk" Apr 14 13:32:51.858362 kubelet[2506]: E0414 13:32:51.854647 2506 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-55d9ccd5cb-vc6bk" Apr 14 13:32:51.858362 kubelet[2506]: E0414 13:32:51.855779 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-55d9ccd5cb-vc6bk_calico-system(8a9902ab-3b55-411b-9c72-16f6b24f69c0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-55d9ccd5cb-vc6bk_calico-system(8a9902ab-3b55-411b-9c72-16f6b24f69c0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-55d9ccd5cb-vc6bk" podUID="8a9902ab-3b55-411b-9c72-16f6b24f69c0" Apr 14 13:32:51.860390 containerd[1454]: time="2026-04-14T13:32:51.857593299Z" level=error msg="Failed to destroy network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.866177 containerd[1454]: time="2026-04-14T13:32:51.865905508Z" level=error msg="encountered an error cleaning up failed sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.872404 containerd[1454]: time="2026-04-14T13:32:51.868726203Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-9f7667bb8-268pt,Uid:10f4fd58-e76c-4b5c-a2bd-c998eae079c3,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.872748 kubelet[2506]: E0414 13:32:51.870340 2506 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:51.872748 kubelet[2506]: E0414 13:32:51.870459 2506 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-9f7667bb8-268pt" Apr 14 13:32:51.872748 kubelet[2506]: E0414 13:32:51.870530 2506 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-9f7667bb8-268pt" Apr 14 13:32:51.872879 kubelet[2506]: E0414 13:32:51.870823 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-9f7667bb8-268pt_calico-system(10f4fd58-e76c-4b5c-a2bd-c998eae079c3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-9f7667bb8-268pt_calico-system(10f4fd58-e76c-4b5c-a2bd-c998eae079c3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-9f7667bb8-268pt" podUID="10f4fd58-e76c-4b5c-a2bd-c998eae079c3" Apr 14 13:32:51.979622 kubelet[2506]: I0414 13:32:51.976441 2506 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:32:52.043326 kubelet[2506]: I0414 13:32:52.004588 2506 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:32:52.056358 containerd[1454]: time="2026-04-14T13:32:52.051144810Z" level=info msg="StopPodSandbox for \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\"" Apr 14 13:32:52.056358 containerd[1454]: time="2026-04-14T13:32:52.053096011Z" level=info msg="Ensure that sandbox 0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8 in task-service has been cleanup successfully" Apr 14 13:32:52.063741 containerd[1454]: time="2026-04-14T13:32:52.059862649Z" level=info msg="StopPodSandbox for \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\"" Apr 14 13:32:52.063741 containerd[1454]: time="2026-04-14T13:32:52.060056975Z" level=info msg="Ensure that sandbox 47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c in task-service has been cleanup successfully" Apr 14 13:32:52.069344 kubelet[2506]: I0414 13:32:52.068339 2506 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:32:52.093609 containerd[1454]: time="2026-04-14T13:32:52.084364133Z" level=info msg="StopPodSandbox for \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\"" Apr 14 13:32:52.093609 containerd[1454]: time="2026-04-14T13:32:52.090005967Z" level=info msg="Ensure that sandbox c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97 in task-service has been cleanup successfully" Apr 14 13:32:52.181095 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71-shm.mount: Deactivated successfully. Apr 14 13:32:52.184074 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78-shm.mount: Deactivated successfully. Apr 14 13:32:52.328324 containerd[1454]: time="2026-04-14T13:32:52.319545259Z" level=error msg="Failed to destroy network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:52.328324 containerd[1454]: time="2026-04-14T13:32:52.321159437Z" level=error msg="encountered an error cleaning up failed sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:52.328324 containerd[1454]: time="2026-04-14T13:32:52.321363954Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56fd9bc5cd-hqzvr,Uid:ada6aff6-b823-4c1b-8ba1-78289ce2a978,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:52.328753 kubelet[2506]: E0414 13:32:52.324547 2506 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:52.328753 kubelet[2506]: E0414 13:32:52.324665 2506 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-56fd9bc5cd-hqzvr" Apr 14 13:32:52.328753 kubelet[2506]: E0414 13:32:52.324714 2506 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-56fd9bc5cd-hqzvr" Apr 14 13:32:52.329066 kubelet[2506]: E0414 13:32:52.325190 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-56fd9bc5cd-hqzvr_calico-system(ada6aff6-b823-4c1b-8ba1-78289ce2a978)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-56fd9bc5cd-hqzvr_calico-system(ada6aff6-b823-4c1b-8ba1-78289ce2a978)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-56fd9bc5cd-hqzvr" podUID="ada6aff6-b823-4c1b-8ba1-78289ce2a978" Apr 14 13:32:52.330663 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f-shm.mount: Deactivated successfully. Apr 14 13:32:52.371027 systemd[1]: run-containerd-runc-k8s.io-bdc6f593c419f26c1f7641ebeaa5a32319685d6d23d132caf7c51c2491f9ae4c-runc.eUc9k5.mount: Deactivated successfully. Apr 14 13:32:52.521713 containerd[1454]: time="2026-04-14T13:32:52.520973692Z" level=error msg="StopPodSandbox for \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\" failed" error="failed to destroy network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:52.529685 kubelet[2506]: E0414 13:32:52.529556 2506 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:32:52.529802 kubelet[2506]: E0414 13:32:52.529697 2506 kuberuntime_manager.go:1881] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8"} Apr 14 13:32:52.529802 kubelet[2506]: E0414 13:32:52.529758 2506 kuberuntime_manager.go:1422] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f6f60262-e827-44cc-9078-cdb0fd928b6c\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 14 13:32:52.530152 kubelet[2506]: E0414 13:32:52.529797 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f6f60262-e827-44cc-9078-cdb0fd928b6c\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-56fd9bc5cd-62gjx" podUID="f6f60262-e827-44cc-9078-cdb0fd928b6c" Apr 14 13:32:52.724741 containerd[1454]: time="2026-04-14T13:32:52.724651871Z" level=error msg="StopPodSandbox for \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\" failed" error="failed to destroy network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:52.726153 kubelet[2506]: E0414 13:32:52.726029 2506 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:32:52.726553 containerd[1454]: time="2026-04-14T13:32:52.726503677Z" level=error msg="StopPodSandbox for \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\" failed" error="failed to destroy network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:52.726942 kubelet[2506]: E0414 13:32:52.726921 2506 kuberuntime_manager.go:1881] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97"} Apr 14 13:32:52.727090 kubelet[2506]: E0414 13:32:52.727056 2506 kuberuntime_manager.go:1422] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 14 13:32:52.727258 kubelet[2506]: E0414 13:32:52.727192 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-5xvdj" podUID="b5b0d33b-1b54-43cc-a4d3-cd788e8eb914" Apr 14 13:32:52.729477 kubelet[2506]: E0414 13:32:52.726809 2506 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:32:52.736875 kubelet[2506]: E0414 13:32:52.736114 2506 kuberuntime_manager.go:1881] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c"} Apr 14 13:32:52.738671 kubelet[2506]: E0414 13:32:52.737908 2506 kuberuntime_manager.go:1422] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"9395319e-7ca7-4248-8723-a075a8a69da5\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 14 13:32:52.741335 kubelet[2506]: E0414 13:32:52.738076 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"9395319e-7ca7-4248-8723-a075a8a69da5\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7d764666f9-km2ks" podUID="9395319e-7ca7-4248-8723-a075a8a69da5" Apr 14 13:32:53.132081 kubelet[2506]: I0414 13:32:53.131913 2506 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:32:53.136296 containerd[1454]: time="2026-04-14T13:32:53.135533106Z" level=info msg="StopPodSandbox for \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\"" Apr 14 13:32:53.139501 containerd[1454]: time="2026-04-14T13:32:53.138846378Z" level=info msg="Ensure that sandbox 5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f in task-service has been cleanup successfully" Apr 14 13:32:53.142938 kubelet[2506]: I0414 13:32:53.142817 2506 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:32:53.145555 containerd[1454]: time="2026-04-14T13:32:53.144962691Z" level=info msg="StopPodSandbox for \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\"" Apr 14 13:32:53.145555 containerd[1454]: time="2026-04-14T13:32:53.145542947Z" level=info msg="Ensure that sandbox b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71 in task-service has been cleanup successfully" Apr 14 13:32:53.212992 kubelet[2506]: I0414 13:32:53.212898 2506 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:32:53.241139 kubelet[2506]: I0414 13:32:53.239971 2506 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:32:53.253724 containerd[1454]: time="2026-04-14T13:32:53.252464582Z" level=info msg="StopPodSandbox for \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\"" Apr 14 13:32:53.260838 containerd[1454]: time="2026-04-14T13:32:53.260718090Z" level=info msg="Ensure that sandbox 0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9 in task-service has been cleanup successfully" Apr 14 13:32:53.272158 containerd[1454]: time="2026-04-14T13:32:53.271798667Z" level=info msg="StopPodSandbox for \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\"" Apr 14 13:32:53.355172 containerd[1454]: time="2026-04-14T13:32:53.354981911Z" level=info msg="Ensure that sandbox 41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78 in task-service has been cleanup successfully" Apr 14 13:32:53.389454 kubelet[2506]: I0414 13:32:53.385865 2506 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:32:53.473331 containerd[1454]: time="2026-04-14T13:32:53.471685771Z" level=info msg="StopPodSandbox for \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\"" Apr 14 13:32:53.482527 containerd[1454]: time="2026-04-14T13:32:53.481455444Z" level=info msg="Ensure that sandbox 6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970 in task-service has been cleanup successfully" Apr 14 13:32:53.551315 containerd[1454]: time="2026-04-14T13:32:53.548336515Z" level=error msg="StopPodSandbox for \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\" failed" error="failed to destroy network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:53.554297 kubelet[2506]: E0414 13:32:53.551673 2506 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:32:53.554297 kubelet[2506]: E0414 13:32:53.551933 2506 kuberuntime_manager.go:1881] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71"} Apr 14 13:32:53.554638 kubelet[2506]: E0414 13:32:53.554161 2506 kuberuntime_manager.go:1422] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"10f4fd58-e76c-4b5c-a2bd-c998eae079c3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 14 13:32:53.554908 kubelet[2506]: E0414 13:32:53.554835 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"10f4fd58-e76c-4b5c-a2bd-c998eae079c3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-9f7667bb8-268pt" podUID="10f4fd58-e76c-4b5c-a2bd-c998eae079c3" Apr 14 13:32:53.561739 containerd[1454]: time="2026-04-14T13:32:53.558312940Z" level=error msg="StopPodSandbox for \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\" failed" error="failed to destroy network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:53.566706 kubelet[2506]: E0414 13:32:53.564680 2506 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:32:53.566706 kubelet[2506]: E0414 13:32:53.565060 2506 kuberuntime_manager.go:1881] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f"} Apr 14 13:32:53.566706 kubelet[2506]: E0414 13:32:53.565143 2506 kuberuntime_manager.go:1422] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ada6aff6-b823-4c1b-8ba1-78289ce2a978\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 14 13:32:53.566706 kubelet[2506]: E0414 13:32:53.565184 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ada6aff6-b823-4c1b-8ba1-78289ce2a978\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-56fd9bc5cd-hqzvr" podUID="ada6aff6-b823-4c1b-8ba1-78289ce2a978" Apr 14 13:32:53.780473 containerd[1454]: time="2026-04-14T13:32:53.780401999Z" level=error msg="StopPodSandbox for \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\" failed" error="failed to destroy network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:53.792733 kubelet[2506]: E0414 13:32:53.792416 2506 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:32:53.797470 kubelet[2506]: E0414 13:32:53.796703 2506 kuberuntime_manager.go:1881] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9"} Apr 14 13:32:53.799704 kubelet[2506]: E0414 13:32:53.799175 2506 kuberuntime_manager.go:1422] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"8d625da5-d2c2-45ba-9200-2567610f3552\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 14 13:32:53.799704 kubelet[2506]: E0414 13:32:53.799449 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"8d625da5-d2c2-45ba-9200-2567610f3552\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7d764666f9-m5vbx" podUID="8d625da5-d2c2-45ba-9200-2567610f3552" Apr 14 13:32:53.888042 containerd[1454]: time="2026-04-14T13:32:53.887852169Z" level=error msg="StopPodSandbox for \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\" failed" error="failed to destroy network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:53.890819 kubelet[2506]: E0414 13:32:53.889536 2506 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:32:53.890819 kubelet[2506]: E0414 13:32:53.889716 2506 kuberuntime_manager.go:1881] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78"} Apr 14 13:32:53.890819 kubelet[2506]: E0414 13:32:53.890583 2506 kuberuntime_manager.go:1422] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 14 13:32:53.891400 kubelet[2506]: E0414 13:32:53.890896 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-55d9ccd5cb-vc6bk" podUID="8a9902ab-3b55-411b-9c72-16f6b24f69c0" Apr 14 13:32:53.951060 containerd[1454]: time="2026-04-14T13:32:53.950963734Z" level=error msg="StopPodSandbox for \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\" failed" error="failed to destroy network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 14 13:32:53.960787 systemd[1]: run-containerd-runc-k8s.io-bdc6f593c419f26c1f7641ebeaa5a32319685d6d23d132caf7c51c2491f9ae4c-runc.I9e1oK.mount: Deactivated successfully. Apr 14 13:32:53.962044 kubelet[2506]: E0414 13:32:53.960915 2506 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:32:53.962044 kubelet[2506]: E0414 13:32:53.961044 2506 kuberuntime_manager.go:1881] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970"} Apr 14 13:32:53.962044 kubelet[2506]: E0414 13:32:53.961652 2506 kuberuntime_manager.go:1422] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"2bca4db2-bc4e-488c-88eb-36317594e234\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 14 13:32:53.962044 kubelet[2506]: E0414 13:32:53.961800 2506 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"2bca4db2-bc4e-488c-88eb-36317594e234\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bb46d7bb5-xwkl7" podUID="2bca4db2-bc4e-488c-88eb-36317594e234" Apr 14 13:32:55.679420 kubelet[2506]: I0414 13:32:55.678949 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-node-ch7dd" podStartSLOduration=8.504605246 podStartE2EDuration="45.678876173s" podCreationTimestamp="2026-04-14 13:32:10 +0000 UTC" firstStartedPulling="2026-04-14 13:32:11.91980763 +0000 UTC m=+21.665814010" lastFinishedPulling="2026-04-14 13:32:49.094078548 +0000 UTC m=+58.840084937" observedRunningTime="2026-04-14 13:32:53.531688797 +0000 UTC m=+63.277695183" watchObservedRunningTime="2026-04-14 13:32:55.678876173 +0000 UTC m=+65.424882565" Apr 14 13:32:55.700188 containerd[1454]: time="2026-04-14T13:32:55.700060315Z" level=info msg="StopPodSandbox for \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\"" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.258 [INFO][4006] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.261 [INFO][4006] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" iface="eth0" netns="/var/run/netns/cni-37c514db-ea6d-ae87-b672-fcd3359660fa" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.262 [INFO][4006] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" iface="eth0" netns="/var/run/netns/cni-37c514db-ea6d-ae87-b672-fcd3359660fa" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.268 [INFO][4006] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" iface="eth0" netns="/var/run/netns/cni-37c514db-ea6d-ae87-b672-fcd3359660fa" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.268 [INFO][4006] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.268 [INFO][4006] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.343 [INFO][4024] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.344 [INFO][4024] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.345 [INFO][4024] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.393 [WARNING][4024] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.393 [INFO][4024] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.466 [INFO][4024] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:32:56.476114 containerd[1454]: 2026-04-14 13:32:56.473 [INFO][4006] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:32:56.476757 containerd[1454]: time="2026-04-14T13:32:56.476582476Z" level=info msg="TearDown network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\" successfully" Apr 14 13:32:56.476757 containerd[1454]: time="2026-04-14T13:32:56.476611491Z" level=info msg="StopPodSandbox for \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\" returns successfully" Apr 14 13:32:56.479103 systemd[1]: run-netns-cni\x2d37c514db\x2dea6d\x2dae87\x2db672\x2dfcd3359660fa.mount: Deactivated successfully. Apr 14 13:32:56.560091 kubelet[2506]: I0414 13:32:56.559948 2506 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-nginx-config\" (UniqueName: \"kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-nginx-config\") pod \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\" (UID: \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\") " Apr 14 13:32:56.560405 kubelet[2506]: I0414 13:32:56.560146 2506 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/projected/8a9902ab-3b55-411b-9c72-16f6b24f69c0-kube-api-access-s4b5j\" (UniqueName: \"kubernetes.io/projected/8a9902ab-3b55-411b-9c72-16f6b24f69c0-kube-api-access-s4b5j\") pod \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\" (UID: \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\") " Apr 14 13:32:56.560405 kubelet[2506]: I0414 13:32:56.560260 2506 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-ca-bundle\") pod \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\" (UID: \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\") " Apr 14 13:32:56.560405 kubelet[2506]: I0414 13:32:56.560358 2506 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/secret/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-backend-key-pair\") pod \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\" (UID: \"8a9902ab-3b55-411b-9c72-16f6b24f69c0\") " Apr 14 13:32:56.560884 kubelet[2506]: I0414 13:32:56.560747 2506 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-nginx-config" pod "8a9902ab-3b55-411b-9c72-16f6b24f69c0" (UID: "8a9902ab-3b55-411b-9c72-16f6b24f69c0"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 14 13:32:56.560884 kubelet[2506]: I0414 13:32:56.560842 2506 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-ca-bundle" pod "8a9902ab-3b55-411b-9c72-16f6b24f69c0" (UID: "8a9902ab-3b55-411b-9c72-16f6b24f69c0"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 14 13:32:56.576764 kubelet[2506]: I0414 13:32:56.576627 2506 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a9902ab-3b55-411b-9c72-16f6b24f69c0-kube-api-access-s4b5j" pod "8a9902ab-3b55-411b-9c72-16f6b24f69c0" (UID: "8a9902ab-3b55-411b-9c72-16f6b24f69c0"). InnerVolumeSpecName "kube-api-access-s4b5j". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 14 13:32:56.577173 kubelet[2506]: I0414 13:32:56.577091 2506 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-backend-key-pair" pod "8a9902ab-3b55-411b-9c72-16f6b24f69c0" (UID: "8a9902ab-3b55-411b-9c72-16f6b24f69c0"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 14 13:32:56.637663 systemd[1]: var-lib-kubelet-pods-8a9902ab\x2d3b55\x2d411b\x2d9c72\x2d16f6b24f69c0-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ds4b5j.mount: Deactivated successfully. Apr 14 13:32:56.661678 systemd[1]: var-lib-kubelet-pods-8a9902ab\x2d3b55\x2d411b\x2d9c72\x2d16f6b24f69c0-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Apr 14 13:32:56.665118 kubelet[2506]: I0414 13:32:56.664372 2506 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-s4b5j\" (UniqueName: \"kubernetes.io/projected/8a9902ab-3b55-411b-9c72-16f6b24f69c0-kube-api-access-s4b5j\") on node \"localhost\" DevicePath \"\"" Apr 14 13:32:56.665118 kubelet[2506]: I0414 13:32:56.664448 2506 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Apr 14 13:32:56.665118 kubelet[2506]: I0414 13:32:56.664458 2506 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/8a9902ab-3b55-411b-9c72-16f6b24f69c0-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Apr 14 13:32:56.665118 kubelet[2506]: I0414 13:32:56.664472 2506 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/8a9902ab-3b55-411b-9c72-16f6b24f69c0-nginx-config\") on node \"localhost\" DevicePath \"\"" Apr 14 13:32:57.505619 systemd[1]: Removed slice kubepods-besteffort-pod8a9902ab_3b55_411b_9c72_16f6b24f69c0.slice - libcontainer container kubepods-besteffort-pod8a9902ab_3b55_411b_9c72_16f6b24f69c0.slice. Apr 14 13:32:58.199387 systemd[1]: Created slice kubepods-besteffort-pod073669a2_a55d_4239_a4a8_960906a2daba.slice - libcontainer container kubepods-besteffort-pod073669a2_a55d_4239_a4a8_960906a2daba.slice. Apr 14 13:32:58.246456 kubelet[2506]: I0414 13:32:58.206983 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/073669a2-a55d-4239-a4a8-960906a2daba-nginx-config\") pod \"whisker-79ddd678c7-cmb4r\" (UID: \"073669a2-a55d-4239-a4a8-960906a2daba\") " pod="calico-system/whisker-79ddd678c7-cmb4r" Apr 14 13:32:58.246456 kubelet[2506]: I0414 13:32:58.207089 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/073669a2-a55d-4239-a4a8-960906a2daba-whisker-ca-bundle\") pod \"whisker-79ddd678c7-cmb4r\" (UID: \"073669a2-a55d-4239-a4a8-960906a2daba\") " pod="calico-system/whisker-79ddd678c7-cmb4r" Apr 14 13:32:58.246456 kubelet[2506]: I0414 13:32:58.207114 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/073669a2-a55d-4239-a4a8-960906a2daba-whisker-backend-key-pair\") pod \"whisker-79ddd678c7-cmb4r\" (UID: \"073669a2-a55d-4239-a4a8-960906a2daba\") " pod="calico-system/whisker-79ddd678c7-cmb4r" Apr 14 13:32:58.246456 kubelet[2506]: I0414 13:32:58.207176 2506 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtsp9\" (UniqueName: \"kubernetes.io/projected/073669a2-a55d-4239-a4a8-960906a2daba-kube-api-access-jtsp9\") pod \"whisker-79ddd678c7-cmb4r\" (UID: \"073669a2-a55d-4239-a4a8-960906a2daba\") " pod="calico-system/whisker-79ddd678c7-cmb4r" Apr 14 13:32:58.508640 kubelet[2506]: I0414 13:32:58.506201 2506 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="8a9902ab-3b55-411b-9c72-16f6b24f69c0" path="/var/lib/kubelet/pods/8a9902ab-3b55-411b-9c72-16f6b24f69c0/volumes" Apr 14 13:32:58.608365 containerd[1454]: time="2026-04-14T13:32:58.607962631Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-79ddd678c7-cmb4r,Uid:073669a2-a55d-4239-a4a8-960906a2daba,Namespace:calico-system,Attempt:0,}" Apr 14 13:33:00.059890 systemd-networkd[1377]: caliee3b3eb5fc4: Link UP Apr 14 13:33:00.060635 systemd-networkd[1377]: caliee3b3eb5fc4: Gained carrier Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:58.778 [ERROR][4038] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:58.957 [INFO][4038] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--79ddd678c7--cmb4r-eth0 whisker-79ddd678c7- calico-system 073669a2-a55d-4239-a4a8-960906a2daba 1042 0 2026-04-14 13:32:57 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:79ddd678c7 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-79ddd678c7-cmb4r eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] caliee3b3eb5fc4 [] [] }} ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Namespace="calico-system" Pod="whisker-79ddd678c7-cmb4r" WorkloadEndpoint="localhost-k8s-whisker--79ddd678c7--cmb4r-" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:58.957 [INFO][4038] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Namespace="calico-system" Pod="whisker-79ddd678c7-cmb4r" WorkloadEndpoint="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.179 [INFO][4051] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" HandleID="k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Workload="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.226 [INFO][4051] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" HandleID="k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Workload="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000326190), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-79ddd678c7-cmb4r", "timestamp":"2026-04-14 13:32:59.179085626 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00017a420)} Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.228 [INFO][4051] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.229 [INFO][4051] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.229 [INFO][4051] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.270 [INFO][4051] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.441 [INFO][4051] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.515 [INFO][4051] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.566 [INFO][4051] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.678 [INFO][4051] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.679 [INFO][4051] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.708 [INFO][4051] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.806 [INFO][4051] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.951 [INFO][4051] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.953 [INFO][4051] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" host="localhost" Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.953 [INFO][4051] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:00.261518 containerd[1454]: 2026-04-14 13:32:59.953 [INFO][4051] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" HandleID="k8s-pod-network.f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Workload="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" Apr 14 13:33:00.267986 containerd[1454]: 2026-04-14 13:32:59.964 [INFO][4038] cni-plugin/k8s.go 418: Populated endpoint ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Namespace="calico-system" Pod="whisker-79ddd678c7-cmb4r" WorkloadEndpoint="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--79ddd678c7--cmb4r-eth0", GenerateName:"whisker-79ddd678c7-", Namespace:"calico-system", SelfLink:"", UID:"073669a2-a55d-4239-a4a8-960906a2daba", ResourceVersion:"1042", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 57, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"79ddd678c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-79ddd678c7-cmb4r", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"caliee3b3eb5fc4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:00.267986 containerd[1454]: 2026-04-14 13:32:59.967 [INFO][4038] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Namespace="calico-system" Pod="whisker-79ddd678c7-cmb4r" WorkloadEndpoint="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" Apr 14 13:33:00.267986 containerd[1454]: 2026-04-14 13:32:59.967 [INFO][4038] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliee3b3eb5fc4 ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Namespace="calico-system" Pod="whisker-79ddd678c7-cmb4r" WorkloadEndpoint="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" Apr 14 13:33:00.267986 containerd[1454]: 2026-04-14 13:33:00.061 [INFO][4038] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Namespace="calico-system" Pod="whisker-79ddd678c7-cmb4r" WorkloadEndpoint="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" Apr 14 13:33:00.267986 containerd[1454]: 2026-04-14 13:33:00.070 [INFO][4038] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Namespace="calico-system" Pod="whisker-79ddd678c7-cmb4r" WorkloadEndpoint="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--79ddd678c7--cmb4r-eth0", GenerateName:"whisker-79ddd678c7-", Namespace:"calico-system", SelfLink:"", UID:"073669a2-a55d-4239-a4a8-960906a2daba", ResourceVersion:"1042", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 57, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"79ddd678c7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d", Pod:"whisker-79ddd678c7-cmb4r", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"caliee3b3eb5fc4", MAC:"76:f8:48:b3:5d:69", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:00.267986 containerd[1454]: 2026-04-14 13:33:00.250 [INFO][4038] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d" Namespace="calico-system" Pod="whisker-79ddd678c7-cmb4r" WorkloadEndpoint="localhost-k8s-whisker--79ddd678c7--cmb4r-eth0" Apr 14 13:33:00.397370 containerd[1454]: time="2026-04-14T13:33:00.394612077Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:33:00.397370 containerd[1454]: time="2026-04-14T13:33:00.394720671Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:33:00.397370 containerd[1454]: time="2026-04-14T13:33:00.394730735Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:00.399127 containerd[1454]: time="2026-04-14T13:33:00.396283621Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:00.466541 systemd[1]: Started cri-containerd-f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d.scope - libcontainer container f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d. Apr 14 13:33:00.501646 systemd-resolved[1379]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 14 13:33:00.581317 containerd[1454]: time="2026-04-14T13:33:00.580336922Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-79ddd678c7-cmb4r,Uid:073669a2-a55d-4239-a4a8-960906a2daba,Namespace:calico-system,Attempt:0,} returns sandbox id \"f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d\"" Apr 14 13:33:00.688925 containerd[1454]: time="2026-04-14T13:33:00.688779646Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\"" Apr 14 13:33:01.186711 systemd-networkd[1377]: caliee3b3eb5fc4: Gained IPv6LL Apr 14 13:33:01.513151 kubelet[2506]: E0414 13:33:01.504155 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:02.201371 kernel: calico-node[4176]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Apr 14 13:33:03.355900 containerd[1454]: time="2026-04-14T13:33:03.355006311Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:03.356636 containerd[1454]: time="2026-04-14T13:33:03.356531481Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.4: active requests=0, bytes read=6039889" Apr 14 13:33:03.368107 containerd[1454]: time="2026-04-14T13:33:03.367955526Z" level=info msg="ImageCreate event name:\"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:03.383358 containerd[1454]: time="2026-04-14T13:33:03.383244537Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:03.387883 containerd[1454]: time="2026-04-14T13:33:03.387839339Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.4\" with image id \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\", size \"7595926\" in 2.699016625s" Apr 14 13:33:03.389419 containerd[1454]: time="2026-04-14T13:33:03.388026045Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\" returns image reference \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\"" Apr 14 13:33:03.464334 containerd[1454]: time="2026-04-14T13:33:03.463636676Z" level=info msg="CreateContainer within sandbox \"f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Apr 14 13:33:03.491294 containerd[1454]: time="2026-04-14T13:33:03.490783367Z" level=info msg="StopPodSandbox for \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\"" Apr 14 13:33:03.501843 containerd[1454]: time="2026-04-14T13:33:03.501685669Z" level=info msg="CreateContainer within sandbox \"f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"8eb1c1499b6f87b66564d65d3e9598fa7621e520435c052b62660b41830d6765\"" Apr 14 13:33:03.537020 containerd[1454]: time="2026-04-14T13:33:03.536951291Z" level=info msg="StartContainer for \"8eb1c1499b6f87b66564d65d3e9598fa7621e520435c052b62660b41830d6765\"" Apr 14 13:33:03.891339 systemd[1]: Started cri-containerd-8eb1c1499b6f87b66564d65d3e9598fa7621e520435c052b62660b41830d6765.scope - libcontainer container 8eb1c1499b6f87b66564d65d3e9598fa7621e520435c052b62660b41830d6765. Apr 14 13:33:04.194957 containerd[1454]: time="2026-04-14T13:33:04.194832669Z" level=info msg="StartContainer for \"8eb1c1499b6f87b66564d65d3e9598fa7621e520435c052b62660b41830d6765\" returns successfully" Apr 14 13:33:04.203051 containerd[1454]: time="2026-04-14T13:33:04.202809318Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\"" Apr 14 13:33:04.413050 systemd-networkd[1377]: vxlan.calico: Link UP Apr 14 13:33:04.413061 systemd-networkd[1377]: vxlan.calico: Gained carrier Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.216 [INFO][4289] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.220 [INFO][4289] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" iface="eth0" netns="/var/run/netns/cni-616eb57a-8910-d02e-91de-8a1ddaf1dc52" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.222 [INFO][4289] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" iface="eth0" netns="/var/run/netns/cni-616eb57a-8910-d02e-91de-8a1ddaf1dc52" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.226 [INFO][4289] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" iface="eth0" netns="/var/run/netns/cni-616eb57a-8910-d02e-91de-8a1ddaf1dc52" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.226 [INFO][4289] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.226 [INFO][4289] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.383 [INFO][4334] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.383 [INFO][4334] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.383 [INFO][4334] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.558 [WARNING][4334] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.559 [INFO][4334] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.641 [INFO][4334] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:04.655757 containerd[1454]: 2026-04-14 13:33:04.647 [INFO][4289] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:33:04.668888 containerd[1454]: time="2026-04-14T13:33:04.656477533Z" level=info msg="TearDown network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\" successfully" Apr 14 13:33:04.668888 containerd[1454]: time="2026-04-14T13:33:04.656549065Z" level=info msg="StopPodSandbox for \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\" returns successfully" Apr 14 13:33:04.662923 systemd[1]: run-netns-cni\x2d616eb57a\x2d8910\x2dd02e\x2d91de\x2d8a1ddaf1dc52.mount: Deactivated successfully. Apr 14 13:33:04.680841 containerd[1454]: time="2026-04-14T13:33:04.680800455Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5xvdj,Uid:b5b0d33b-1b54-43cc-a4d3-cd788e8eb914,Namespace:calico-system,Attempt:1,}" Apr 14 13:33:05.508519 containerd[1454]: time="2026-04-14T13:33:05.506058536Z" level=info msg="StopPodSandbox for \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\"" Apr 14 13:33:05.513396 containerd[1454]: time="2026-04-14T13:33:05.513322040Z" level=info msg="StopPodSandbox for \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\"" Apr 14 13:33:05.519662 containerd[1454]: time="2026-04-14T13:33:05.519429164Z" level=info msg="StopPodSandbox for \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\"" Apr 14 13:33:06.247674 systemd-networkd[1377]: vxlan.calico: Gained IPv6LL Apr 14 13:33:06.780845 containerd[1454]: time="2026-04-14T13:33:06.780587861Z" level=info msg="StopPodSandbox for \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\"" Apr 14 13:33:06.995434 kubelet[2506]: E0414 13:33:06.992740 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:07.133173 containerd[1454]: time="2026-04-14T13:33:07.132950609Z" level=info msg="StopPodSandbox for \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\"" Apr 14 13:33:07.224045 systemd-networkd[1377]: cali648a174a2cb: Link UP Apr 14 13:33:07.269183 systemd-networkd[1377]: cali648a174a2cb: Gained carrier Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:06.450 [INFO][4426] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:06.450 [INFO][4426] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" iface="eth0" netns="/var/run/netns/cni-3ce494ff-4798-c1f0-b2eb-bc12ef3388da" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:06.451 [INFO][4426] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" iface="eth0" netns="/var/run/netns/cni-3ce494ff-4798-c1f0-b2eb-bc12ef3388da" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:06.451 [INFO][4426] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" iface="eth0" netns="/var/run/netns/cni-3ce494ff-4798-c1f0-b2eb-bc12ef3388da" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:06.451 [INFO][4426] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:06.451 [INFO][4426] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:06.959 [INFO][4489] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:06.961 [INFO][4489] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:07.151 [INFO][4489] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:07.239 [WARNING][4489] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:07.240 [INFO][4489] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:07.395 [INFO][4489] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:07.427276 containerd[1454]: 2026-04-14 13:33:07.417 [INFO][4426] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:33:07.427276 containerd[1454]: time="2026-04-14T13:33:07.425246826Z" level=info msg="TearDown network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\" successfully" Apr 14 13:33:07.427276 containerd[1454]: time="2026-04-14T13:33:07.425276937Z" level=info msg="StopPodSandbox for \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\" returns successfully" Apr 14 13:33:07.441533 kubelet[2506]: E0414 13:33:07.437964 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:07.443597 containerd[1454]: time="2026-04-14T13:33:07.443488049Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-km2ks,Uid:9395319e-7ca7-4248-8723-a075a8a69da5,Namespace:kube-system,Attempt:1,}" Apr 14 13:33:07.451354 systemd[1]: run-netns-cni\x2d3ce494ff\x2d4798\x2dc1f0\x2db2eb\x2dbc12ef3388da.mount: Deactivated successfully. Apr 14 13:33:07.573299 containerd[1454]: time="2026-04-14T13:33:07.554504167Z" level=info msg="StopPodSandbox for \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\"" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.081 [INFO][4357] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--5xvdj-eth0 csi-node-driver- calico-system b5b0d33b-1b54-43cc-a4d3-cd788e8eb914 1065 0 2026-04-14 13:32:10 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:589b8b8d94 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-5xvdj eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali648a174a2cb [] [] }} ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Namespace="calico-system" Pod="csi-node-driver-5xvdj" WorkloadEndpoint="localhost-k8s-csi--node--driver--5xvdj-" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.082 [INFO][4357] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Namespace="calico-system" Pod="csi-node-driver-5xvdj" WorkloadEndpoint="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.321 [INFO][4389] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" HandleID="k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.469 [INFO][4389] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" HandleID="k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000592c60), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-5xvdj", "timestamp":"2026-04-14 13:33:05.321889663 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000198c60)} Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.469 [INFO][4389] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.469 [INFO][4389] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.469 [INFO][4389] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.557 [INFO][4389] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:05.893 [INFO][4389] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:06.135 [INFO][4389] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:06.165 [INFO][4389] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:06.355 [INFO][4389] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:06.355 [INFO][4389] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:06.497 [INFO][4389] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041 Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:06.843 [INFO][4389] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:07.145 [INFO][4389] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:07.150 [INFO][4389] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" host="localhost" Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:07.151 [INFO][4389] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:07.808433 containerd[1454]: 2026-04-14 13:33:07.153 [INFO][4389] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" HandleID="k8s-pod-network.542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:07.821000 containerd[1454]: 2026-04-14 13:33:07.161 [INFO][4357] cni-plugin/k8s.go 418: Populated endpoint ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Namespace="calico-system" Pod="csi-node-driver-5xvdj" WorkloadEndpoint="localhost-k8s-csi--node--driver--5xvdj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--5xvdj-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914", ResourceVersion:"1065", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"589b8b8d94", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-5xvdj", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali648a174a2cb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:07.821000 containerd[1454]: 2026-04-14 13:33:07.164 [INFO][4357] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Namespace="calico-system" Pod="csi-node-driver-5xvdj" WorkloadEndpoint="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:07.821000 containerd[1454]: 2026-04-14 13:33:07.164 [INFO][4357] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali648a174a2cb ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Namespace="calico-system" Pod="csi-node-driver-5xvdj" WorkloadEndpoint="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:07.821000 containerd[1454]: 2026-04-14 13:33:07.268 [INFO][4357] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Namespace="calico-system" Pod="csi-node-driver-5xvdj" WorkloadEndpoint="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:07.821000 containerd[1454]: 2026-04-14 13:33:07.297 [INFO][4357] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Namespace="calico-system" Pod="csi-node-driver-5xvdj" WorkloadEndpoint="localhost-k8s-csi--node--driver--5xvdj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--5xvdj-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914", ResourceVersion:"1065", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"589b8b8d94", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041", Pod:"csi-node-driver-5xvdj", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali648a174a2cb", MAC:"0e:bb:4b:2f:2b:f0", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:07.821000 containerd[1454]: 2026-04-14 13:33:07.673 [INFO][4357] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041" Namespace="calico-system" Pod="csi-node-driver-5xvdj" WorkloadEndpoint="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:06.393 [INFO][4447] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:06.393 [INFO][4447] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" iface="eth0" netns="/var/run/netns/cni-27367552-b5c8-6e68-076d-014c61253eb2" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:06.394 [INFO][4447] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" iface="eth0" netns="/var/run/netns/cni-27367552-b5c8-6e68-076d-014c61253eb2" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:06.394 [INFO][4447] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" iface="eth0" netns="/var/run/netns/cni-27367552-b5c8-6e68-076d-014c61253eb2" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:06.394 [INFO][4447] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:06.394 [INFO][4447] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:06.968 [INFO][4487] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:06.997 [INFO][4487] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:07.398 [INFO][4487] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:07.634 [WARNING][4487] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:07.643 [INFO][4487] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:07.830 [INFO][4487] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:07.870860 containerd[1454]: 2026-04-14 13:33:07.850 [INFO][4447] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:33:07.885545 containerd[1454]: time="2026-04-14T13:33:07.874644315Z" level=info msg="TearDown network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\" successfully" Apr 14 13:33:07.885545 containerd[1454]: time="2026-04-14T13:33:07.878352536Z" level=info msg="StopPodSandbox for \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\" returns successfully" Apr 14 13:33:07.890098 systemd[1]: run-netns-cni\x2d27367552\x2db5c8\x2d6e68\x2d076d\x2d014c61253eb2.mount: Deactivated successfully. Apr 14 13:33:07.976524 containerd[1454]: time="2026-04-14T13:33:07.970670763Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56fd9bc5cd-hqzvr,Uid:ada6aff6-b823-4c1b-8ba1-78289ce2a978,Namespace:calico-system,Attempt:1,}" Apr 14 13:33:08.253309 containerd[1454]: time="2026-04-14T13:33:08.247500137Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:33:08.253309 containerd[1454]: time="2026-04-14T13:33:08.247617027Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:33:08.253309 containerd[1454]: time="2026-04-14T13:33:08.247629300Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:08.253309 containerd[1454]: time="2026-04-14T13:33:08.247787999Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:06.679 [INFO][4448] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:06.680 [INFO][4448] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" iface="eth0" netns="/var/run/netns/cni-f8da8a4d-895b-d7cf-7efc-2f97b68083bf" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:06.682 [INFO][4448] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" iface="eth0" netns="/var/run/netns/cni-f8da8a4d-895b-d7cf-7efc-2f97b68083bf" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:06.685 [INFO][4448] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" iface="eth0" netns="/var/run/netns/cni-f8da8a4d-895b-d7cf-7efc-2f97b68083bf" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:06.685 [INFO][4448] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:06.685 [INFO][4448] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:07.405 [INFO][4501] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:07.406 [INFO][4501] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:07.858 [INFO][4501] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:08.202 [WARNING][4501] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:08.203 [INFO][4501] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:08.276 [INFO][4501] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:08.411273 containerd[1454]: 2026-04-14 13:33:08.295 [INFO][4448] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:33:08.411273 containerd[1454]: time="2026-04-14T13:33:08.411166726Z" level=info msg="TearDown network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\" successfully" Apr 14 13:33:08.411273 containerd[1454]: time="2026-04-14T13:33:08.411210605Z" level=info msg="StopPodSandbox for \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\" returns successfully" Apr 14 13:33:08.447793 systemd[1]: run-netns-cni\x2df8da8a4d\x2d895b\x2dd7cf\x2d7efc\x2d2f97b68083bf.mount: Deactivated successfully. Apr 14 13:33:08.474601 containerd[1454]: time="2026-04-14T13:33:08.472557298Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bb46d7bb5-xwkl7,Uid:2bca4db2-bc4e-488c-88eb-36317594e234,Namespace:calico-system,Attempt:1,}" Apr 14 13:33:08.477884 systemd[1]: Started cri-containerd-542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041.scope - libcontainer container 542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041. Apr 14 13:33:09.064417 systemd-resolved[1379]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 14 13:33:09.244912 systemd-networkd[1377]: cali648a174a2cb: Gained IPv6LL Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:08.405 [INFO][4520] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:08.423 [INFO][4520] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" iface="eth0" netns="/var/run/netns/cni-ce71664f-bd4e-42cb-c17b-6660385a59ad" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:08.468 [INFO][4520] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" iface="eth0" netns="/var/run/netns/cni-ce71664f-bd4e-42cb-c17b-6660385a59ad" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:08.492 [INFO][4520] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" iface="eth0" netns="/var/run/netns/cni-ce71664f-bd4e-42cb-c17b-6660385a59ad" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:08.500 [INFO][4520] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:08.500 [INFO][4520] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:09.053 [INFO][4637] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:09.055 [INFO][4637] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:09.055 [INFO][4637] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:09.266 [WARNING][4637] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:09.267 [INFO][4637] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:09.500 [INFO][4637] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:09.657587 containerd[1454]: 2026-04-14 13:33:09.636 [INFO][4520] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:09.673594 containerd[1454]: time="2026-04-14T13:33:09.673395327Z" level=info msg="TearDown network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\" successfully" Apr 14 13:33:09.673594 containerd[1454]: time="2026-04-14T13:33:09.673549032Z" level=info msg="StopPodSandbox for \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\" returns successfully" Apr 14 13:33:09.677605 systemd[1]: run-netns-cni\x2dce71664f\x2dbd4e\x2d42cb\x2dc17b\x2d6660385a59ad.mount: Deactivated successfully. Apr 14 13:33:09.740746 containerd[1454]: time="2026-04-14T13:33:09.739122077Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5xvdj,Uid:b5b0d33b-1b54-43cc-a4d3-cd788e8eb914,Namespace:calico-system,Attempt:1,} returns sandbox id \"542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041\"" Apr 14 13:33:09.752853 containerd[1454]: time="2026-04-14T13:33:09.750394449Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-9f7667bb8-268pt,Uid:10f4fd58-e76c-4b5c-a2bd-c998eae079c3,Namespace:calico-system,Attempt:1,}" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.282 [INFO][4542] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.369 [INFO][4542] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" iface="eth0" netns="/var/run/netns/cni-7007711b-6f50-e042-0b74-7a6dfc11d2db" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.369 [INFO][4542] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" iface="eth0" netns="/var/run/netns/cni-7007711b-6f50-e042-0b74-7a6dfc11d2db" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.389 [INFO][4542] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" iface="eth0" netns="/var/run/netns/cni-7007711b-6f50-e042-0b74-7a6dfc11d2db" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.389 [INFO][4542] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.389 [INFO][4542] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.921 [INFO][4677] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.926 [INFO][4677] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:09.930 [INFO][4677] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:10.143 [WARNING][4677] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:10.143 [INFO][4677] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:10.477 [INFO][4677] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:10.519439 containerd[1454]: 2026-04-14 13:33:10.513 [INFO][4542] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:10.530425 containerd[1454]: time="2026-04-14T13:33:10.527070247Z" level=info msg="TearDown network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\" successfully" Apr 14 13:33:10.530425 containerd[1454]: time="2026-04-14T13:33:10.527183402Z" level=info msg="StopPodSandbox for \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\" returns successfully" Apr 14 13:33:10.538823 systemd[1]: run-netns-cni\x2d7007711b\x2d6f50\x2de042\x2d0b74\x2d7a6dfc11d2db.mount: Deactivated successfully. Apr 14 13:33:10.679406 kubelet[2506]: E0414 13:33:10.674805 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:10.686166 containerd[1454]: time="2026-04-14T13:33:10.685551674Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-m5vbx,Uid:8d625da5-d2c2-45ba-9200-2567610f3552,Namespace:kube-system,Attempt:1,}" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:09.586 [INFO][4570] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:09.617 [INFO][4570] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" iface="eth0" netns="/var/run/netns/cni-c89c8713-0ee2-14e9-ee11-2662018f78a5" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:09.621 [INFO][4570] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" iface="eth0" netns="/var/run/netns/cni-c89c8713-0ee2-14e9-ee11-2662018f78a5" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:09.629 [INFO][4570] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" iface="eth0" netns="/var/run/netns/cni-c89c8713-0ee2-14e9-ee11-2662018f78a5" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:09.629 [INFO][4570] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:09.629 [INFO][4570] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:10.075 [INFO][4690] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:10.077 [INFO][4690] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:10.478 [INFO][4690] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:10.729 [WARNING][4690] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:10.729 [INFO][4690] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:10.904 [INFO][4690] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:10.928580 containerd[1454]: 2026-04-14 13:33:10.918 [INFO][4570] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:10.949382 containerd[1454]: time="2026-04-14T13:33:10.946355795Z" level=info msg="TearDown network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\" successfully" Apr 14 13:33:10.949382 containerd[1454]: time="2026-04-14T13:33:10.946420835Z" level=info msg="StopPodSandbox for \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\" returns successfully" Apr 14 13:33:10.978679 systemd[1]: run-netns-cni\x2dc89c8713\x2d0ee2\x2d14e9\x2dee11\x2d2662018f78a5.mount: Deactivated successfully. Apr 14 13:33:10.987068 containerd[1454]: time="2026-04-14T13:33:10.983722424Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56fd9bc5cd-62gjx,Uid:f6f60262-e827-44cc-9078-cdb0fd928b6c,Namespace:calico-system,Attempt:1,}" Apr 14 13:33:12.480202 systemd-networkd[1377]: cali1510ebdd2f6: Link UP Apr 14 13:33:12.493950 systemd-networkd[1377]: cali1510ebdd2f6: Gained carrier Apr 14 13:33:12.758095 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1516727006.mount: Deactivated successfully. Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:09.073 [INFO][4575] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7d764666f9--km2ks-eth0 coredns-7d764666f9- kube-system 9395319e-7ca7-4248-8723-a075a8a69da5 1076 0 2026-04-14 13:31:54 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7d764666f9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7d764666f9-km2ks eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali1510ebdd2f6 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Namespace="kube-system" Pod="coredns-7d764666f9-km2ks" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--km2ks-" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:09.156 [INFO][4575] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Namespace="kube-system" Pod="coredns-7d764666f9-km2ks" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:10.081 [INFO][4694] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" HandleID="k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:10.555 [INFO][4694] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" HandleID="k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002fde70), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7d764666f9-km2ks", "timestamp":"2026-04-14 13:33:10.081554324 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00037a2c0)} Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:10.558 [INFO][4694] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:10.904 [INFO][4694] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:10.904 [INFO][4694] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:11.258 [INFO][4694] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:11.495 [INFO][4694] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:11.694 [INFO][4694] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:11.759 [INFO][4694] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:11.856 [INFO][4694] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:11.862 [INFO][4694] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:11.964 [INFO][4694] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:12.190 [INFO][4694] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:12.378 [INFO][4694] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:12.380 [INFO][4694] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" host="localhost" Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:12.380 [INFO][4694] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:12.831085 containerd[1454]: 2026-04-14 13:33:12.380 [INFO][4694] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" HandleID="k8s-pod-network.e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:12.832811 containerd[1454]: 2026-04-14 13:33:12.394 [INFO][4575] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Namespace="kube-system" Pod="coredns-7d764666f9-km2ks" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--km2ks-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--km2ks-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"9395319e-7ca7-4248-8723-a075a8a69da5", ResourceVersion:"1076", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 31, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7d764666f9-km2ks", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1510ebdd2f6", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:12.832811 containerd[1454]: 2026-04-14 13:33:12.395 [INFO][4575] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Namespace="kube-system" Pod="coredns-7d764666f9-km2ks" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:12.832811 containerd[1454]: 2026-04-14 13:33:12.396 [INFO][4575] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1510ebdd2f6 ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Namespace="kube-system" Pod="coredns-7d764666f9-km2ks" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:12.832811 containerd[1454]: 2026-04-14 13:33:12.498 [INFO][4575] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Namespace="kube-system" Pod="coredns-7d764666f9-km2ks" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:12.832811 containerd[1454]: 2026-04-14 13:33:12.534 [INFO][4575] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Namespace="kube-system" Pod="coredns-7d764666f9-km2ks" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--km2ks-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--km2ks-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"9395319e-7ca7-4248-8723-a075a8a69da5", ResourceVersion:"1076", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 31, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff", Pod:"coredns-7d764666f9-km2ks", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1510ebdd2f6", MAC:"66:fe:c5:b9:de:d3", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:12.832811 containerd[1454]: 2026-04-14 13:33:12.736 [INFO][4575] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff" Namespace="kube-system" Pod="coredns-7d764666f9-km2ks" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:33:12.942965 containerd[1454]: time="2026-04-14T13:33:12.942892534Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:12.990056 containerd[1454]: time="2026-04-14T13:33:12.989872084Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.4: active requests=0, bytes read=17609475" Apr 14 13:33:13.005847 containerd[1454]: time="2026-04-14T13:33:13.005532757Z" level=info msg="ImageCreate event name:\"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:13.082907 containerd[1454]: time="2026-04-14T13:33:13.080876507Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:13.097108 containerd[1454]: time="2026-04-14T13:33:13.094962726Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" with image id \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\", size \"17609305\" in 8.892115369s" Apr 14 13:33:13.097108 containerd[1454]: time="2026-04-14T13:33:13.095025111Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" returns image reference \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\"" Apr 14 13:33:13.191287 containerd[1454]: time="2026-04-14T13:33:13.190967098Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:33:13.191287 containerd[1454]: time="2026-04-14T13:33:13.191120641Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:33:13.191287 containerd[1454]: time="2026-04-14T13:33:13.191134952Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:13.191651 containerd[1454]: time="2026-04-14T13:33:13.191278332Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:13.228674 containerd[1454]: time="2026-04-14T13:33:13.227754866Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\"" Apr 14 13:33:13.269835 containerd[1454]: time="2026-04-14T13:33:13.269753157Z" level=info msg="CreateContainer within sandbox \"f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Apr 14 13:33:13.279623 systemd[1]: Started cri-containerd-e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff.scope - libcontainer container e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff. Apr 14 13:33:13.381006 systemd-resolved[1379]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 14 13:33:13.398943 containerd[1454]: time="2026-04-14T13:33:13.398625925Z" level=info msg="CreateContainer within sandbox \"f86e514375a52cde5c3e48482087eb59f9bf509079dd25afa8a337e2b22c917d\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"6be659b27704b2cbbd0a95bfe9e9d3b4942b89f3b50abe118c09ff1ad8deedcc\"" Apr 14 13:33:13.424524 containerd[1454]: time="2026-04-14T13:33:13.423373219Z" level=info msg="StartContainer for \"6be659b27704b2cbbd0a95bfe9e9d3b4942b89f3b50abe118c09ff1ad8deedcc\"" Apr 14 13:33:13.467355 containerd[1454]: time="2026-04-14T13:33:13.466262022Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-km2ks,Uid:9395319e-7ca7-4248-8723-a075a8a69da5,Namespace:kube-system,Attempt:1,} returns sandbox id \"e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff\"" Apr 14 13:33:13.574972 kubelet[2506]: E0414 13:33:13.574860 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:13.619714 containerd[1454]: time="2026-04-14T13:33:13.619643746Z" level=info msg="CreateContainer within sandbox \"e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 14 13:33:13.662538 containerd[1454]: time="2026-04-14T13:33:13.662469555Z" level=info msg="CreateContainer within sandbox \"e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"619e16847302c242b0cc144ea5a09a8286227d363410c624f07a754c6ef7eed7\"" Apr 14 13:33:13.686324 containerd[1454]: time="2026-04-14T13:33:13.685269111Z" level=info msg="StartContainer for \"619e16847302c242b0cc144ea5a09a8286227d363410c624f07a754c6ef7eed7\"" Apr 14 13:33:13.817828 systemd[1]: Started cri-containerd-6be659b27704b2cbbd0a95bfe9e9d3b4942b89f3b50abe118c09ff1ad8deedcc.scope - libcontainer container 6be659b27704b2cbbd0a95bfe9e9d3b4942b89f3b50abe118c09ff1ad8deedcc. Apr 14 13:33:13.897719 systemd-networkd[1377]: calidbd249e75f3: Link UP Apr 14 13:33:13.903321 systemd-networkd[1377]: calidbd249e75f3: Gained carrier Apr 14 13:33:13.921489 systemd-networkd[1377]: cali1510ebdd2f6: Gained IPv6LL Apr 14 13:33:14.064652 systemd[1]: Started cri-containerd-619e16847302c242b0cc144ea5a09a8286227d363410c624f07a754c6ef7eed7.scope - libcontainer container 619e16847302c242b0cc144ea5a09a8286227d363410c624f07a754c6ef7eed7. Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:09.473 [INFO][4613] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0 calico-apiserver-56fd9bc5cd- calico-system ada6aff6-b823-4c1b-8ba1-78289ce2a978 1074 0 2026-04-14 13:32:08 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:56fd9bc5cd projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-56fd9bc5cd-hqzvr eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] calidbd249e75f3 [] [] }} ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-hqzvr" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:09.585 [INFO][4613] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-hqzvr" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:10.247 [INFO][4705] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" HandleID="k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:10.603 [INFO][4705] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" HandleID="k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ee030), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-apiserver-56fd9bc5cd-hqzvr", "timestamp":"2026-04-14 13:33:10.247563198 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0000ff1e0)} Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:10.605 [INFO][4705] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:12.382 [INFO][4705] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:12.382 [INFO][4705] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:12.483 [INFO][4705] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:12.719 [INFO][4705] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.193 [INFO][4705] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.269 [INFO][4705] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.306 [INFO][4705] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.306 [INFO][4705] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.389 [INFO][4705] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0 Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.440 [INFO][4705] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.846 [INFO][4705] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.846 [INFO][4705] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" host="localhost" Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.852 [INFO][4705] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:14.257367 containerd[1454]: 2026-04-14 13:33:13.854 [INFO][4705] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" HandleID="k8s-pod-network.6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:14.258805 containerd[1454]: 2026-04-14 13:33:13.861 [INFO][4613] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-hqzvr" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0", GenerateName:"calico-apiserver-56fd9bc5cd-", Namespace:"calico-system", SelfLink:"", UID:"ada6aff6-b823-4c1b-8ba1-78289ce2a978", ResourceVersion:"1074", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56fd9bc5cd", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-56fd9bc5cd-hqzvr", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calidbd249e75f3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:14.258805 containerd[1454]: 2026-04-14 13:33:13.872 [INFO][4613] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-hqzvr" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:14.258805 containerd[1454]: 2026-04-14 13:33:13.873 [INFO][4613] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidbd249e75f3 ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-hqzvr" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:14.258805 containerd[1454]: 2026-04-14 13:33:13.922 [INFO][4613] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-hqzvr" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:14.258805 containerd[1454]: 2026-04-14 13:33:13.960 [INFO][4613] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-hqzvr" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0", GenerateName:"calico-apiserver-56fd9bc5cd-", Namespace:"calico-system", SelfLink:"", UID:"ada6aff6-b823-4c1b-8ba1-78289ce2a978", ResourceVersion:"1074", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56fd9bc5cd", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0", Pod:"calico-apiserver-56fd9bc5cd-hqzvr", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calidbd249e75f3", MAC:"de:a8:1e:1b:cf:06", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:14.258805 containerd[1454]: 2026-04-14 13:33:14.245 [INFO][4613] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-hqzvr" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:33:14.528440 containerd[1454]: time="2026-04-14T13:33:14.524585133Z" level=info msg="StartContainer for \"619e16847302c242b0cc144ea5a09a8286227d363410c624f07a754c6ef7eed7\" returns successfully" Apr 14 13:33:14.661484 containerd[1454]: time="2026-04-14T13:33:14.595197760Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:33:14.661484 containerd[1454]: time="2026-04-14T13:33:14.603716133Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:33:14.661484 containerd[1454]: time="2026-04-14T13:33:14.603742574Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:14.661484 containerd[1454]: time="2026-04-14T13:33:14.603895832Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:14.767439 containerd[1454]: time="2026-04-14T13:33:14.767345875Z" level=info msg="StartContainer for \"6be659b27704b2cbbd0a95bfe9e9d3b4942b89f3b50abe118c09ff1ad8deedcc\" returns successfully" Apr 14 13:33:14.896642 systemd[1]: Started cri-containerd-6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0.scope - libcontainer container 6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0. Apr 14 13:33:15.134560 kubelet[2506]: E0414 13:33:15.128941 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:15.388032 systemd-networkd[1377]: calidbd249e75f3: Gained IPv6LL Apr 14 13:33:15.535421 kubelet[2506]: I0414 13:33:15.533053 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/coredns-7d764666f9-km2ks" podStartSLOduration=81.532979031 podStartE2EDuration="1m21.532979031s" podCreationTimestamp="2026-04-14 13:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-14 13:33:15.530261582 +0000 UTC m=+85.276267996" watchObservedRunningTime="2026-04-14 13:33:15.532979031 +0000 UTC m=+85.278985417" Apr 14 13:33:15.600659 systemd-resolved[1379]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 14 13:33:15.760246 systemd-networkd[1377]: calib713ab3c088: Link UP Apr 14 13:33:15.764071 systemd-networkd[1377]: calib713ab3c088: Gained carrier Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:09.624 [INFO][4652] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0 calico-kube-controllers-5bb46d7bb5- calico-system 2bca4db2-bc4e-488c-88eb-36317594e234 1077 0 2026-04-14 13:32:11 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5bb46d7bb5 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-5bb46d7bb5-xwkl7 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calib713ab3c088 [] [] }} ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Namespace="calico-system" Pod="calico-kube-controllers-5bb46d7bb5-xwkl7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:09.624 [INFO][4652] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Namespace="calico-system" Pod="calico-kube-controllers-5bb46d7bb5-xwkl7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:10.447 [INFO][4710] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" HandleID="k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:10.706 [INFO][4710] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" HandleID="k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000388610), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-5bb46d7bb5-xwkl7", "timestamp":"2026-04-14 13:33:10.44720535 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00038e840)} Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:10.706 [INFO][4710] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:13.849 [INFO][4710] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:13.852 [INFO][4710] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:14.066 [INFO][4710] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:14.295 [INFO][4710] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:14.855 [INFO][4710] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:14.937 [INFO][4710] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:15.245 [INFO][4710] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:15.245 [INFO][4710] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:15.292 [INFO][4710] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:15.482 [INFO][4710] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:15.681 [INFO][4710] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:15.682 [INFO][4710] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" host="localhost" Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:15.686 [INFO][4710] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:15.959206 containerd[1454]: 2026-04-14 13:33:15.688 [INFO][4710] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" HandleID="k8s-pod-network.5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:15.962058 containerd[1454]: 2026-04-14 13:33:15.749 [INFO][4652] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Namespace="calico-system" Pod="calico-kube-controllers-5bb46d7bb5-xwkl7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0", GenerateName:"calico-kube-controllers-5bb46d7bb5-", Namespace:"calico-system", SelfLink:"", UID:"2bca4db2-bc4e-488c-88eb-36317594e234", ResourceVersion:"1077", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bb46d7bb5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-5bb46d7bb5-xwkl7", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib713ab3c088", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:15.962058 containerd[1454]: 2026-04-14 13:33:15.749 [INFO][4652] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Namespace="calico-system" Pod="calico-kube-controllers-5bb46d7bb5-xwkl7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:15.962058 containerd[1454]: 2026-04-14 13:33:15.749 [INFO][4652] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib713ab3c088 ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Namespace="calico-system" Pod="calico-kube-controllers-5bb46d7bb5-xwkl7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:15.962058 containerd[1454]: 2026-04-14 13:33:15.765 [INFO][4652] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Namespace="calico-system" Pod="calico-kube-controllers-5bb46d7bb5-xwkl7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:15.962058 containerd[1454]: 2026-04-14 13:33:15.765 [INFO][4652] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Namespace="calico-system" Pod="calico-kube-controllers-5bb46d7bb5-xwkl7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0", GenerateName:"calico-kube-controllers-5bb46d7bb5-", Namespace:"calico-system", SelfLink:"", UID:"2bca4db2-bc4e-488c-88eb-36317594e234", ResourceVersion:"1077", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bb46d7bb5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e", Pod:"calico-kube-controllers-5bb46d7bb5-xwkl7", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib713ab3c088", MAC:"2a:c3:61:90:9c:fe", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:15.962058 containerd[1454]: 2026-04-14 13:33:15.940 [INFO][4652] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e" Namespace="calico-system" Pod="calico-kube-controllers-5bb46d7bb5-xwkl7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:16.149384 containerd[1454]: time="2026-04-14T13:33:16.143928290Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:33:16.162008 containerd[1454]: time="2026-04-14T13:33:16.154932937Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:33:16.171558 containerd[1454]: time="2026-04-14T13:33:16.169140313Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:16.192207 containerd[1454]: time="2026-04-14T13:33:16.190515861Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:16.290191 kubelet[2506]: E0414 13:33:16.288805 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:16.500177 systemd[1]: run-containerd-runc-k8s.io-5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e-runc.blCPPt.mount: Deactivated successfully. Apr 14 13:33:16.703046 containerd[1454]: time="2026-04-14T13:33:16.702943081Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56fd9bc5cd-hqzvr,Uid:ada6aff6-b823-4c1b-8ba1-78289ce2a978,Namespace:calico-system,Attempt:1,} returns sandbox id \"6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0\"" Apr 14 13:33:16.779824 systemd[1]: Started cri-containerd-5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e.scope - libcontainer container 5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e. Apr 14 13:33:16.794979 systemd[1]: Started sshd@7-10.0.0.8:22-10.0.0.1:34884.service - OpenSSH per-connection server daemon (10.0.0.1:34884). Apr 14 13:33:17.088108 systemd-resolved[1379]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 14 13:33:17.120384 sshd[5048]: Accepted publickey for core from 10.0.0.1 port 34884 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:33:17.138934 sshd[5048]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:33:17.206645 systemd-logind[1442]: New session 8 of user core. Apr 14 13:33:17.268396 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 14 13:33:17.407965 kubelet[2506]: I0414 13:33:17.400615 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/whisker-79ddd678c7-cmb4r" podStartSLOduration=7.913573607 podStartE2EDuration="20.400596542s" podCreationTimestamp="2026-04-14 13:32:57 +0000 UTC" firstStartedPulling="2026-04-14 13:33:00.674985468 +0000 UTC m=+70.420991852" lastFinishedPulling="2026-04-14 13:33:13.162008391 +0000 UTC m=+82.908014787" observedRunningTime="2026-04-14 13:33:17.350877049 +0000 UTC m=+87.096883439" watchObservedRunningTime="2026-04-14 13:33:17.400596542 +0000 UTC m=+87.146602922" Apr 14 13:33:17.655876 systemd-networkd[1377]: calib713ab3c088: Gained IPv6LL Apr 14 13:33:17.688301 kubelet[2506]: E0414 13:33:17.681759 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:17.688301 kubelet[2506]: E0414 13:33:17.681943 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:17.933856 containerd[1454]: time="2026-04-14T13:33:17.933055249Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bb46d7bb5-xwkl7,Uid:2bca4db2-bc4e-488c-88eb-36317594e234,Namespace:calico-system,Attempt:1,} returns sandbox id \"5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e\"" Apr 14 13:33:18.397031 containerd[1454]: time="2026-04-14T13:33:18.396653206Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:18.398618 containerd[1454]: time="2026-04-14T13:33:18.398174670Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.4: active requests=0, bytes read=8792502" Apr 14 13:33:18.403293 containerd[1454]: time="2026-04-14T13:33:18.403205789Z" level=info msg="ImageCreate event name:\"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:18.483259 containerd[1454]: time="2026-04-14T13:33:18.483110833Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:18.496980 containerd[1454]: time="2026-04-14T13:33:18.496793396Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.4\" with image id \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\", size \"10348547\" in 5.268913565s" Apr 14 13:33:18.497315 containerd[1454]: time="2026-04-14T13:33:18.497028357Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\" returns image reference \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\"" Apr 14 13:33:18.512694 containerd[1454]: time="2026-04-14T13:33:18.512616612Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Apr 14 13:33:18.539966 containerd[1454]: time="2026-04-14T13:33:18.539893047Z" level=info msg="CreateContainer within sandbox \"542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Apr 14 13:33:18.732018 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3572655959.mount: Deactivated successfully. Apr 14 13:33:18.741601 containerd[1454]: time="2026-04-14T13:33:18.740986522Z" level=info msg="CreateContainer within sandbox \"542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"d485e5cf77bf0399f3daf7f5b35937f066f6e26c81a542c8a0efb11e20abe4d0\"" Apr 14 13:33:18.742943 containerd[1454]: time="2026-04-14T13:33:18.742866787Z" level=info msg="StartContainer for \"d485e5cf77bf0399f3daf7f5b35937f066f6e26c81a542c8a0efb11e20abe4d0\"" Apr 14 13:33:19.138869 systemd[1]: Started cri-containerd-d485e5cf77bf0399f3daf7f5b35937f066f6e26c81a542c8a0efb11e20abe4d0.scope - libcontainer container d485e5cf77bf0399f3daf7f5b35937f066f6e26c81a542c8a0efb11e20abe4d0. Apr 14 13:33:19.360114 systemd-networkd[1377]: cali2e05b9f506b: Link UP Apr 14 13:33:19.361624 systemd-networkd[1377]: cali2e05b9f506b: Gained carrier Apr 14 13:33:19.667995 sshd[5048]: pam_unix(sshd:session): session closed for user core Apr 14 13:33:19.698663 systemd[1]: sshd@7-10.0.0.8:22-10.0.0.1:34884.service: Deactivated successfully. Apr 14 13:33:19.710803 systemd[1]: session-8.scope: Deactivated successfully. Apr 14 13:33:19.726018 systemd-logind[1442]: Session 8 logged out. Waiting for processes to exit. Apr 14 13:33:19.734434 systemd-logind[1442]: Removed session 8. Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:11.144 [INFO][4718] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--9f7667bb8--268pt-eth0 goldmane-9f7667bb8- calico-system 10f4fd58-e76c-4b5c-a2bd-c998eae079c3 1090 0 2026-04-14 13:32:09 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:9f7667bb8 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-9f7667bb8-268pt eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali2e05b9f506b [] [] }} ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Namespace="calico-system" Pod="goldmane-9f7667bb8-268pt" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--268pt-" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:11.153 [INFO][4718] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Namespace="calico-system" Pod="goldmane-9f7667bb8-268pt" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:11.707 [INFO][4769] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" HandleID="k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:11.850 [INFO][4769] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" HandleID="k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000376990), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-9f7667bb8-268pt", "timestamp":"2026-04-14 13:33:11.707459784 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000367080)} Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:11.851 [INFO][4769] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:15.682 [INFO][4769] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:15.682 [INFO][4769] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:15.798 [INFO][4769] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:16.201 [INFO][4769] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:16.819 [INFO][4769] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:17.290 [INFO][4769] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:17.419 [INFO][4769] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:17.422 [INFO][4769] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:17.689 [INFO][4769] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432 Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:18.132 [INFO][4769] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:19.331 [INFO][4769] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:19.332 [INFO][4769] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" host="localhost" Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:19.335 [INFO][4769] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:19.767494 containerd[1454]: 2026-04-14 13:33:19.335 [INFO][4769] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" HandleID="k8s-pod-network.0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:19.792596 containerd[1454]: 2026-04-14 13:33:19.344 [INFO][4718] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Namespace="calico-system" Pod="goldmane-9f7667bb8-268pt" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--9f7667bb8--268pt-eth0", GenerateName:"goldmane-9f7667bb8-", Namespace:"calico-system", SelfLink:"", UID:"10f4fd58-e76c-4b5c-a2bd-c998eae079c3", ResourceVersion:"1090", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"9f7667bb8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-9f7667bb8-268pt", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali2e05b9f506b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:19.792596 containerd[1454]: 2026-04-14 13:33:19.344 [INFO][4718] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Namespace="calico-system" Pod="goldmane-9f7667bb8-268pt" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:19.792596 containerd[1454]: 2026-04-14 13:33:19.344 [INFO][4718] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2e05b9f506b ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Namespace="calico-system" Pod="goldmane-9f7667bb8-268pt" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:19.792596 containerd[1454]: 2026-04-14 13:33:19.358 [INFO][4718] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Namespace="calico-system" Pod="goldmane-9f7667bb8-268pt" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:19.792596 containerd[1454]: 2026-04-14 13:33:19.368 [INFO][4718] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Namespace="calico-system" Pod="goldmane-9f7667bb8-268pt" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--9f7667bb8--268pt-eth0", GenerateName:"goldmane-9f7667bb8-", Namespace:"calico-system", SelfLink:"", UID:"10f4fd58-e76c-4b5c-a2bd-c998eae079c3", ResourceVersion:"1090", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"9f7667bb8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432", Pod:"goldmane-9f7667bb8-268pt", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali2e05b9f506b", MAC:"6e:3e:19:14:a5:c1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:19.792596 containerd[1454]: 2026-04-14 13:33:19.761 [INFO][4718] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432" Namespace="calico-system" Pod="goldmane-9f7667bb8-268pt" WorkloadEndpoint="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:20.126708 containerd[1454]: time="2026-04-14T13:33:20.124592850Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:33:20.128276 containerd[1454]: time="2026-04-14T13:33:20.127997047Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:33:20.128276 containerd[1454]: time="2026-04-14T13:33:20.128069570Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:20.128449 containerd[1454]: time="2026-04-14T13:33:20.128212986Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:20.300088 systemd[1]: Started cri-containerd-0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432.scope - libcontainer container 0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432. Apr 14 13:33:20.643582 containerd[1454]: time="2026-04-14T13:33:20.643484301Z" level=info msg="StartContainer for \"d485e5cf77bf0399f3daf7f5b35937f066f6e26c81a542c8a0efb11e20abe4d0\" returns successfully" Apr 14 13:33:20.762787 systemd-networkd[1377]: cali2e05b9f506b: Gained IPv6LL Apr 14 13:33:21.188517 systemd-resolved[1379]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 14 13:33:21.408610 containerd[1454]: time="2026-04-14T13:33:21.408453278Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-9f7667bb8-268pt,Uid:10f4fd58-e76c-4b5c-a2bd-c998eae079c3,Namespace:calico-system,Attempt:1,} returns sandbox id \"0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432\"" Apr 14 13:33:22.477014 systemd-networkd[1377]: cali6900a99c3cd: Link UP Apr 14 13:33:22.478505 systemd-networkd[1377]: cali6900a99c3cd: Gained carrier Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:11.524 [INFO][4741] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7d764666f9--m5vbx-eth0 coredns-7d764666f9- kube-system 8d625da5-d2c2-45ba-9200-2567610f3552 1091 0 2026-04-14 13:31:54 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7d764666f9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7d764666f9-m5vbx eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali6900a99c3cd [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Namespace="kube-system" Pod="coredns-7d764666f9-m5vbx" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--m5vbx-" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:11.532 [INFO][4741] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Namespace="kube-system" Pod="coredns-7d764666f9-m5vbx" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:11.951 [INFO][4780] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" HandleID="k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:12.142 [INFO][4780] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" HandleID="k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000334740), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7d764666f9-m5vbx", "timestamp":"2026-04-14 13:33:11.951081981 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00032da20)} Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:12.160 [INFO][4780] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:19.336 [INFO][4780] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:19.337 [INFO][4780] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:19.731 [INFO][4780] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:20.520 [INFO][4780] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:21.455 [INFO][4780] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:21.590 [INFO][4780] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:21.879 [INFO][4780] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:21.880 [INFO][4780] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:21.929 [INFO][4780] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:22.082 [INFO][4780] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:22.386 [INFO][4780] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:22.386 [INFO][4780] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" host="localhost" Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:22.386 [INFO][4780] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:23.053433 containerd[1454]: 2026-04-14 13:33:22.386 [INFO][4780] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" HandleID="k8s-pod-network.f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:23.059899 containerd[1454]: 2026-04-14 13:33:22.397 [INFO][4741] cni-plugin/k8s.go 418: Populated endpoint ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Namespace="kube-system" Pod="coredns-7d764666f9-m5vbx" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--m5vbx-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"8d625da5-d2c2-45ba-9200-2567610f3552", ResourceVersion:"1091", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 31, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7d764666f9-m5vbx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6900a99c3cd", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:23.059899 containerd[1454]: 2026-04-14 13:33:22.399 [INFO][4741] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Namespace="kube-system" Pod="coredns-7d764666f9-m5vbx" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:23.059899 containerd[1454]: 2026-04-14 13:33:22.452 [INFO][4741] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6900a99c3cd ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Namespace="kube-system" Pod="coredns-7d764666f9-m5vbx" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:23.059899 containerd[1454]: 2026-04-14 13:33:22.477 [INFO][4741] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Namespace="kube-system" Pod="coredns-7d764666f9-m5vbx" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:23.059899 containerd[1454]: 2026-04-14 13:33:22.525 [INFO][4741] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Namespace="kube-system" Pod="coredns-7d764666f9-m5vbx" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--m5vbx-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"8d625da5-d2c2-45ba-9200-2567610f3552", ResourceVersion:"1091", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 31, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db", Pod:"coredns-7d764666f9-m5vbx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6900a99c3cd", MAC:"62:1b:6d:23:5d:60", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:23.059899 containerd[1454]: 2026-04-14 13:33:23.036 [INFO][4741] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db" Namespace="kube-system" Pod="coredns-7d764666f9-m5vbx" WorkloadEndpoint="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:23.347358 containerd[1454]: time="2026-04-14T13:33:23.339108916Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:33:23.347358 containerd[1454]: time="2026-04-14T13:33:23.339908720Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:33:23.347358 containerd[1454]: time="2026-04-14T13:33:23.339928600Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:23.347358 containerd[1454]: time="2026-04-14T13:33:23.340099845Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:23.532727 systemd[1]: Started cri-containerd-f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db.scope - libcontainer container f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db. Apr 14 13:33:23.759075 systemd-resolved[1379]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 14 13:33:24.104816 containerd[1454]: time="2026-04-14T13:33:24.104503303Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-m5vbx,Uid:8d625da5-d2c2-45ba-9200-2567610f3552,Namespace:kube-system,Attempt:1,} returns sandbox id \"f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db\"" Apr 14 13:33:24.125430 kubelet[2506]: E0414 13:33:24.123660 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:24.348735 containerd[1454]: time="2026-04-14T13:33:24.348023141Z" level=info msg="CreateContainer within sandbox \"f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 14 13:33:24.452959 systemd-networkd[1377]: cali6900a99c3cd: Gained IPv6LL Apr 14 13:33:24.546463 kubelet[2506]: E0414 13:33:24.542541 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:24.574272 containerd[1454]: time="2026-04-14T13:33:24.567924706Z" level=info msg="CreateContainer within sandbox \"f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8436abda11e112ec49fe7690a0b33a98c94daa7e67b2f0207025861e8dfd421c\"" Apr 14 13:33:24.575427 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount250023791.mount: Deactivated successfully. Apr 14 13:33:24.658971 containerd[1454]: time="2026-04-14T13:33:24.658508069Z" level=info msg="StartContainer for \"8436abda11e112ec49fe7690a0b33a98c94daa7e67b2f0207025861e8dfd421c\"" Apr 14 13:33:24.700626 systemd[1]: Started sshd@8-10.0.0.8:22-10.0.0.1:54130.service - OpenSSH per-connection server daemon (10.0.0.1:54130). Apr 14 13:33:25.165922 systemd[1]: run-containerd-runc-k8s.io-8436abda11e112ec49fe7690a0b33a98c94daa7e67b2f0207025861e8dfd421c-runc.hxUymB.mount: Deactivated successfully. Apr 14 13:33:25.170582 sshd[5248]: Accepted publickey for core from 10.0.0.1 port 54130 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:33:25.180407 sshd[5248]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:33:25.205616 systemd[1]: Started cri-containerd-8436abda11e112ec49fe7690a0b33a98c94daa7e67b2f0207025861e8dfd421c.scope - libcontainer container 8436abda11e112ec49fe7690a0b33a98c94daa7e67b2f0207025861e8dfd421c. Apr 14 13:33:25.224452 systemd-logind[1442]: New session 9 of user core. Apr 14 13:33:25.224605 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 14 13:33:25.265027 systemd-networkd[1377]: calibc0079e2f02: Link UP Apr 14 13:33:25.268501 systemd-networkd[1377]: calibc0079e2f02: Gained carrier Apr 14 13:33:25.468888 containerd[1454]: time="2026-04-14T13:33:25.468715614Z" level=info msg="StartContainer for \"8436abda11e112ec49fe7690a0b33a98c94daa7e67b2f0207025861e8dfd421c\" returns successfully" Apr 14 13:33:25.894975 kubelet[2506]: E0414 13:33:25.890761 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:11.698 [INFO][4753] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0 calico-apiserver-56fd9bc5cd- calico-system f6f60262-e827-44cc-9078-cdb0fd928b6c 1092 0 2026-04-14 13:32:08 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:56fd9bc5cd projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-56fd9bc5cd-62gjx eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] calibc0079e2f02 [] [] }} ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-62gjx" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:11.698 [INFO][4753] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-62gjx" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:12.110 [INFO][4788] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" HandleID="k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:12.229 [INFO][4788] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" HandleID="k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ee0a0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-apiserver-56fd9bc5cd-62gjx", "timestamp":"2026-04-14 13:33:12.110976027 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0000fe6e0)} Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:12.230 [INFO][4788] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:22.387 [INFO][4788] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:22.387 [INFO][4788] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:22.694 [INFO][4788] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:22.995 [INFO][4788] ipam/ipam.go 409: Looking up existing affinities for host host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:23.382 [INFO][4788] ipam/ipam.go 526: Trying affinity for 192.168.88.128/26 host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:23.803 [INFO][4788] ipam/ipam.go 160: Attempting to load block cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:24.124 [INFO][4788] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:24.137 [INFO][4788] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:24.473 [INFO][4788] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:24.580 [INFO][4788] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:25.200 [INFO][4788] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:25.200 [INFO][4788] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" host="localhost" Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:25.201 [INFO][4788] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:25.960033 containerd[1454]: 2026-04-14 13:33:25.201 [INFO][4788] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" HandleID="k8s-pod-network.0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:25.960803 containerd[1454]: 2026-04-14 13:33:25.206 [INFO][4753] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-62gjx" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0", GenerateName:"calico-apiserver-56fd9bc5cd-", Namespace:"calico-system", SelfLink:"", UID:"f6f60262-e827-44cc-9078-cdb0fd928b6c", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56fd9bc5cd", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-56fd9bc5cd-62gjx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calibc0079e2f02", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:25.960803 containerd[1454]: 2026-04-14 13:33:25.206 [INFO][4753] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-62gjx" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:25.960803 containerd[1454]: 2026-04-14 13:33:25.206 [INFO][4753] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calibc0079e2f02 ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-62gjx" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:25.960803 containerd[1454]: 2026-04-14 13:33:25.273 [INFO][4753] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-62gjx" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:25.960803 containerd[1454]: 2026-04-14 13:33:25.274 [INFO][4753] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-62gjx" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0", GenerateName:"calico-apiserver-56fd9bc5cd-", Namespace:"calico-system", SelfLink:"", UID:"f6f60262-e827-44cc-9078-cdb0fd928b6c", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56fd9bc5cd", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f", Pod:"calico-apiserver-56fd9bc5cd-62gjx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calibc0079e2f02", MAC:"96:ae:fc:13:6a:42", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:25.960803 containerd[1454]: 2026-04-14 13:33:25.845 [INFO][4753] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f" Namespace="calico-system" Pod="calico-apiserver-56fd9bc5cd-62gjx" WorkloadEndpoint="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:26.258011 containerd[1454]: time="2026-04-14T13:33:26.256322395Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 14 13:33:26.258011 containerd[1454]: time="2026-04-14T13:33:26.256378515Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 14 13:33:26.258011 containerd[1454]: time="2026-04-14T13:33:26.256389990Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:26.258011 containerd[1454]: time="2026-04-14T13:33:26.256458724Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 14 13:33:26.434541 systemd[1]: Started cri-containerd-0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f.scope - libcontainer container 0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f. Apr 14 13:33:26.640562 systemd-resolved[1379]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 14 13:33:26.748821 systemd-networkd[1377]: calibc0079e2f02: Gained IPv6LL Apr 14 13:33:26.804461 sshd[5248]: pam_unix(sshd:session): session closed for user core Apr 14 13:33:26.837129 systemd[1]: sshd@8-10.0.0.8:22-10.0.0.1:54130.service: Deactivated successfully. Apr 14 13:33:26.856635 systemd[1]: session-9.scope: Deactivated successfully. Apr 14 13:33:26.879319 systemd-logind[1442]: Session 9 logged out. Waiting for processes to exit. Apr 14 13:33:26.887703 systemd-logind[1442]: Removed session 9. Apr 14 13:33:26.989169 kubelet[2506]: E0414 13:33:26.989077 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:27.032309 containerd[1454]: time="2026-04-14T13:33:27.031599330Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56fd9bc5cd-62gjx,Uid:f6f60262-e827-44cc-9078-cdb0fd928b6c,Namespace:calico-system,Attempt:1,} returns sandbox id \"0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f\"" Apr 14 13:33:27.699386 kubelet[2506]: E0414 13:33:27.696491 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:28.273681 kubelet[2506]: I0414 13:33:28.272957 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/coredns-7d764666f9-m5vbx" podStartSLOduration=94.272887475 podStartE2EDuration="1m34.272887475s" podCreationTimestamp="2026-04-14 13:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-14 13:33:27.479918088 +0000 UTC m=+97.225924473" watchObservedRunningTime="2026-04-14 13:33:28.272887475 +0000 UTC m=+98.018893857" Apr 14 13:33:31.833790 systemd[1]: Started sshd@9-10.0.0.8:22-10.0.0.1:49202.service - OpenSSH per-connection server daemon (10.0.0.1:49202). Apr 14 13:33:32.021872 sshd[5392]: Accepted publickey for core from 10.0.0.1 port 49202 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:33:32.024741 sshd[5392]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:33:32.071156 systemd-logind[1442]: New session 10 of user core. Apr 14 13:33:32.081970 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 14 13:33:33.464120 containerd[1454]: time="2026-04-14T13:33:33.464064237Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:33.468756 containerd[1454]: time="2026-04-14T13:33:33.468687808Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=48415780" Apr 14 13:33:33.474055 containerd[1454]: time="2026-04-14T13:33:33.473947242Z" level=info msg="ImageCreate event name:\"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:33.481716 containerd[1454]: time="2026-04-14T13:33:33.481613215Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:33.488640 containerd[1454]: time="2026-04-14T13:33:33.487484997Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"49971841\" in 14.974806942s" Apr 14 13:33:33.488640 containerd[1454]: time="2026-04-14T13:33:33.487566075Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\"" Apr 14 13:33:33.555731 containerd[1454]: time="2026-04-14T13:33:33.555498860Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\"" Apr 14 13:33:33.601852 containerd[1454]: time="2026-04-14T13:33:33.601701233Z" level=info msg="CreateContainer within sandbox \"6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 14 13:33:33.714093 sshd[5392]: pam_unix(sshd:session): session closed for user core Apr 14 13:33:33.717089 containerd[1454]: time="2026-04-14T13:33:33.715802497Z" level=info msg="CreateContainer within sandbox \"6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"4eb2e75a988bb688e80d63c6338416e694403475c32e62e6c3e7a9aa064f9d75\"" Apr 14 13:33:33.722733 containerd[1454]: time="2026-04-14T13:33:33.722545768Z" level=info msg="StartContainer for \"4eb2e75a988bb688e80d63c6338416e694403475c32e62e6c3e7a9aa064f9d75\"" Apr 14 13:33:33.726907 systemd[1]: sshd@9-10.0.0.8:22-10.0.0.1:49202.service: Deactivated successfully. Apr 14 13:33:33.729010 systemd[1]: session-10.scope: Deactivated successfully. Apr 14 13:33:33.749470 systemd-logind[1442]: Session 10 logged out. Waiting for processes to exit. Apr 14 13:33:33.761752 systemd-logind[1442]: Removed session 10. Apr 14 13:33:33.911040 systemd[1]: Started cri-containerd-4eb2e75a988bb688e80d63c6338416e694403475c32e62e6c3e7a9aa064f9d75.scope - libcontainer container 4eb2e75a988bb688e80d63c6338416e694403475c32e62e6c3e7a9aa064f9d75. Apr 14 13:33:34.200755 containerd[1454]: time="2026-04-14T13:33:34.200547240Z" level=info msg="StartContainer for \"4eb2e75a988bb688e80d63c6338416e694403475c32e62e6c3e7a9aa064f9d75\" returns successfully" Apr 14 13:33:34.900846 kubelet[2506]: I0414 13:33:34.900071 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-apiserver-56fd9bc5cd-hqzvr" podStartSLOduration=70.280986696 podStartE2EDuration="1m26.899986971s" podCreationTimestamp="2026-04-14 13:32:08 +0000 UTC" firstStartedPulling="2026-04-14 13:33:16.935461056 +0000 UTC m=+86.681467445" lastFinishedPulling="2026-04-14 13:33:33.554461336 +0000 UTC m=+103.300467720" observedRunningTime="2026-04-14 13:33:34.891584203 +0000 UTC m=+104.637590589" watchObservedRunningTime="2026-04-14 13:33:34.899986971 +0000 UTC m=+104.645993357" Apr 14 13:33:36.996394 kubelet[2506]: E0414 13:33:36.992558 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:33:38.772477 systemd[1]: Started sshd@10-10.0.0.8:22-10.0.0.1:49210.service - OpenSSH per-connection server daemon (10.0.0.1:49210). Apr 14 13:33:39.043172 sshd[5467]: Accepted publickey for core from 10.0.0.1 port 49210 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:33:39.037934 sshd[5467]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:33:39.057933 systemd-logind[1442]: New session 11 of user core. Apr 14 13:33:39.063528 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 14 13:33:40.238667 sshd[5467]: pam_unix(sshd:session): session closed for user core Apr 14 13:33:40.248394 systemd[1]: sshd@10-10.0.0.8:22-10.0.0.1:49210.service: Deactivated successfully. Apr 14 13:33:40.265886 systemd[1]: session-11.scope: Deactivated successfully. Apr 14 13:33:40.273840 systemd-logind[1442]: Session 11 logged out. Waiting for processes to exit. Apr 14 13:33:40.276549 systemd-logind[1442]: Removed session 11. Apr 14 13:33:43.920498 containerd[1454]: time="2026-04-14T13:33:43.920138504Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:43.925267 containerd[1454]: time="2026-04-14T13:33:43.924839042Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.4: active requests=0, bytes read=52406348" Apr 14 13:33:43.929926 containerd[1454]: time="2026-04-14T13:33:43.929750328Z" level=info msg="ImageCreate event name:\"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:43.937037 containerd[1454]: time="2026-04-14T13:33:43.936836866Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:43.937716 containerd[1454]: time="2026-04-14T13:33:43.937694147Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" with image id \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\", size \"53962361\" in 10.382030602s" Apr 14 13:33:43.937871 containerd[1454]: time="2026-04-14T13:33:43.937860136Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" returns image reference \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\"" Apr 14 13:33:43.951760 containerd[1454]: time="2026-04-14T13:33:43.951726695Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\"" Apr 14 13:33:44.096022 containerd[1454]: time="2026-04-14T13:33:44.095703836Z" level=info msg="CreateContainer within sandbox \"5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Apr 14 13:33:44.192746 containerd[1454]: time="2026-04-14T13:33:44.192549081Z" level=info msg="CreateContainer within sandbox \"5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"d4b71d3668f315326a69e2c511d2ce9b7e8f9b7128b92db6c73f0de3adb515f5\"" Apr 14 13:33:44.206358 containerd[1454]: time="2026-04-14T13:33:44.204539050Z" level=info msg="StartContainer for \"d4b71d3668f315326a69e2c511d2ce9b7e8f9b7128b92db6c73f0de3adb515f5\"" Apr 14 13:33:44.468831 systemd[1]: Started cri-containerd-d4b71d3668f315326a69e2c511d2ce9b7e8f9b7128b92db6c73f0de3adb515f5.scope - libcontainer container d4b71d3668f315326a69e2c511d2ce9b7e8f9b7128b92db6c73f0de3adb515f5. Apr 14 13:33:44.858119 containerd[1454]: time="2026-04-14T13:33:44.856544930Z" level=info msg="StartContainer for \"d4b71d3668f315326a69e2c511d2ce9b7e8f9b7128b92db6c73f0de3adb515f5\" returns successfully" Apr 14 13:33:45.279810 systemd[1]: Started sshd@11-10.0.0.8:22-10.0.0.1:55240.service - OpenSSH per-connection server daemon (10.0.0.1:55240). Apr 14 13:33:45.424625 kubelet[2506]: I0414 13:33:45.424427 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5bb46d7bb5-xwkl7" podStartSLOduration=68.413542807 podStartE2EDuration="1m34.424215485s" podCreationTimestamp="2026-04-14 13:32:11 +0000 UTC" firstStartedPulling="2026-04-14 13:33:17.940641482 +0000 UTC m=+87.686647862" lastFinishedPulling="2026-04-14 13:33:43.951314159 +0000 UTC m=+113.697320540" observedRunningTime="2026-04-14 13:33:45.387118006 +0000 UTC m=+115.133124397" watchObservedRunningTime="2026-04-14 13:33:45.424215485 +0000 UTC m=+115.170221869" Apr 14 13:33:45.575001 sshd[5532]: Accepted publickey for core from 10.0.0.1 port 55240 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:33:45.579343 sshd[5532]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:33:45.621181 systemd-logind[1442]: New session 12 of user core. Apr 14 13:33:45.627031 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 14 13:33:47.207039 sshd[5532]: pam_unix(sshd:session): session closed for user core Apr 14 13:33:47.267421 systemd[1]: sshd@11-10.0.0.8:22-10.0.0.1:55240.service: Deactivated successfully. Apr 14 13:33:47.281577 systemd[1]: session-12.scope: Deactivated successfully. Apr 14 13:33:47.301078 systemd-logind[1442]: Session 12 logged out. Waiting for processes to exit. Apr 14 13:33:47.341885 systemd-logind[1442]: Removed session 12. Apr 14 13:33:47.963121 containerd[1454]: time="2026-04-14T13:33:47.962636031Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4: active requests=0, bytes read=14704317" Apr 14 13:33:47.964493 containerd[1454]: time="2026-04-14T13:33:47.963259461Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:47.973702 containerd[1454]: time="2026-04-14T13:33:47.969806514Z" level=info msg="ImageCreate event name:\"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:47.982750 containerd[1454]: time="2026-04-14T13:33:47.982659469Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:48.084447 containerd[1454]: time="2026-04-14T13:33:48.081861622Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" with image id \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\", size \"16260314\" in 4.129839467s" Apr 14 13:33:48.084447 containerd[1454]: time="2026-04-14T13:33:48.082169604Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" returns image reference \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\"" Apr 14 13:33:48.104139 containerd[1454]: time="2026-04-14T13:33:48.103869822Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\"" Apr 14 13:33:48.136069 containerd[1454]: time="2026-04-14T13:33:48.135903699Z" level=info msg="CreateContainer within sandbox \"542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Apr 14 13:33:48.216462 containerd[1454]: time="2026-04-14T13:33:48.214681141Z" level=info msg="CreateContainer within sandbox \"542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"2bb904c012db271ab87316c12926496c5661864e6c62ce4aa00edd0efca6c751\"" Apr 14 13:33:48.237570 containerd[1454]: time="2026-04-14T13:33:48.227482099Z" level=info msg="StartContainer for \"2bb904c012db271ab87316c12926496c5661864e6c62ce4aa00edd0efca6c751\"" Apr 14 13:33:48.370581 systemd[1]: Started cri-containerd-2bb904c012db271ab87316c12926496c5661864e6c62ce4aa00edd0efca6c751.scope - libcontainer container 2bb904c012db271ab87316c12926496c5661864e6c62ce4aa00edd0efca6c751. Apr 14 13:33:48.699509 containerd[1454]: time="2026-04-14T13:33:48.699267451Z" level=info msg="StartContainer for \"2bb904c012db271ab87316c12926496c5661864e6c62ce4aa00edd0efca6c751\" returns successfully" Apr 14 13:33:49.274191 kubelet[2506]: I0414 13:33:49.207449 2506 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Apr 14 13:33:49.281761 kubelet[2506]: I0414 13:33:49.241505 2506 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Apr 14 13:33:50.288290 kubelet[2506]: I0414 13:33:50.288098 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/csi-node-driver-5xvdj" podStartSLOduration=62.04026247 podStartE2EDuration="1m40.288048152s" podCreationTimestamp="2026-04-14 13:32:10 +0000 UTC" firstStartedPulling="2026-04-14 13:33:09.850783897 +0000 UTC m=+79.596790280" lastFinishedPulling="2026-04-14 13:33:48.098569569 +0000 UTC m=+117.844575962" observedRunningTime="2026-04-14 13:33:50.279586836 +0000 UTC m=+120.025593233" watchObservedRunningTime="2026-04-14 13:33:50.288048152 +0000 UTC m=+120.034054532" Apr 14 13:33:50.659200 containerd[1454]: time="2026-04-14T13:33:50.659166591Z" level=info msg="StopPodSandbox for \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\"" Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.439 [WARNING][5641] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0", GenerateName:"calico-apiserver-56fd9bc5cd-", Namespace:"calico-system", SelfLink:"", UID:"f6f60262-e827-44cc-9078-cdb0fd928b6c", ResourceVersion:"1216", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56fd9bc5cd", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f", Pod:"calico-apiserver-56fd9bc5cd-62gjx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calibc0079e2f02", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.454 [INFO][5641] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.455 [INFO][5641] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" iface="eth0" netns="" Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.455 [INFO][5641] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.455 [INFO][5641] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.634 [INFO][5650] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.634 [INFO][5650] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.634 [INFO][5650] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.790 [WARNING][5650] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.790 [INFO][5650] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.812 [INFO][5650] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:51.837733 containerd[1454]: 2026-04-14 13:33:51.825 [INFO][5641] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:51.892326 containerd[1454]: time="2026-04-14T13:33:51.892261357Z" level=info msg="TearDown network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\" successfully" Apr 14 13:33:51.892326 containerd[1454]: time="2026-04-14T13:33:51.892311817Z" level=info msg="StopPodSandbox for \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\" returns successfully" Apr 14 13:33:51.984916 containerd[1454]: time="2026-04-14T13:33:51.984616509Z" level=info msg="RemovePodSandbox for \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\"" Apr 14 13:33:51.990887 containerd[1454]: time="2026-04-14T13:33:51.990737253Z" level=info msg="Forcibly stopping sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\"" Apr 14 13:33:52.287195 systemd[1]: Started sshd@12-10.0.0.8:22-10.0.0.1:57296.service - OpenSSH per-connection server daemon (10.0.0.1:57296). Apr 14 13:33:52.470572 sshd[5675]: Accepted publickey for core from 10.0.0.1 port 57296 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:33:52.495703 sshd[5675]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:33:52.578054 systemd-logind[1442]: New session 13 of user core. Apr 14 13:33:52.588821 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:52.872 [WARNING][5668] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0", GenerateName:"calico-apiserver-56fd9bc5cd-", Namespace:"calico-system", SelfLink:"", UID:"f6f60262-e827-44cc-9078-cdb0fd928b6c", ResourceVersion:"1216", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56fd9bc5cd", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f", Pod:"calico-apiserver-56fd9bc5cd-62gjx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calibc0079e2f02", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:52.875 [INFO][5668] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:52.882 [INFO][5668] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" iface="eth0" netns="" Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:52.886 [INFO][5668] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:52.887 [INFO][5668] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:53.073 [INFO][5684] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:53.077 [INFO][5684] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:53.078 [INFO][5684] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:53.426 [WARNING][5684] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:53.427 [INFO][5684] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" HandleID="k8s-pod-network.0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--62gjx-eth0" Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:53.599 [INFO][5684] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:53.652294 containerd[1454]: 2026-04-14 13:33:53.605 [INFO][5668] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8" Apr 14 13:33:53.655552 containerd[1454]: time="2026-04-14T13:33:53.652966075Z" level=info msg="TearDown network for sandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\" successfully" Apr 14 13:33:53.825825 containerd[1454]: time="2026-04-14T13:33:53.818209627Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 14 13:33:53.825825 containerd[1454]: time="2026-04-14T13:33:53.819024988Z" level=info msg="RemovePodSandbox \"0bfe2f4e4e546973d5df759069ca777552b98ede582d9c29b88591d3ef3b66f8\" returns successfully" Apr 14 13:33:53.978150 containerd[1454]: time="2026-04-14T13:33:53.976372941Z" level=info msg="StopPodSandbox for \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\"" Apr 14 13:33:54.553457 sshd[5675]: pam_unix(sshd:session): session closed for user core Apr 14 13:33:54.570905 systemd-logind[1442]: Session 13 logged out. Waiting for processes to exit. Apr 14 13:33:54.574395 systemd[1]: sshd@12-10.0.0.8:22-10.0.0.1:57296.service: Deactivated successfully. Apr 14 13:33:54.576041 systemd[1]: session-13.scope: Deactivated successfully. Apr 14 13:33:54.591461 systemd-logind[1442]: Removed session 13. Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.528 [WARNING][5726] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--m5vbx-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"8d625da5-d2c2-45ba-9200-2567610f3552", ResourceVersion:"1273", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 31, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db", Pod:"coredns-7d764666f9-m5vbx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6900a99c3cd", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.531 [INFO][5726] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.531 [INFO][5726] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" iface="eth0" netns="" Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.532 [INFO][5726] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.534 [INFO][5726] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.872 [INFO][5735] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.873 [INFO][5735] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.873 [INFO][5735] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.969 [WARNING][5735] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.969 [INFO][5735] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.993 [INFO][5735] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:54.999885 containerd[1454]: 2026-04-14 13:33:54.995 [INFO][5726] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:54.999885 containerd[1454]: time="2026-04-14T13:33:54.999784277Z" level=info msg="TearDown network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\" successfully" Apr 14 13:33:54.999885 containerd[1454]: time="2026-04-14T13:33:54.999813933Z" level=info msg="StopPodSandbox for \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\" returns successfully" Apr 14 13:33:55.009279 containerd[1454]: time="2026-04-14T13:33:55.009190683Z" level=info msg="RemovePodSandbox for \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\"" Apr 14 13:33:55.009279 containerd[1454]: time="2026-04-14T13:33:55.009292533Z" level=info msg="Forcibly stopping sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\"" Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:55.624 [WARNING][5759] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--m5vbx-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"8d625da5-d2c2-45ba-9200-2567610f3552", ResourceVersion:"1273", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 31, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f110578a5bb0519b93c66eb329813a6af2fe3ed5323ad539c7af35e0671d42db", Pod:"coredns-7d764666f9-m5vbx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6900a99c3cd", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:55.632 [INFO][5759] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:55.636 [INFO][5759] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" iface="eth0" netns="" Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:55.636 [INFO][5759] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:55.636 [INFO][5759] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:55.887 [INFO][5769] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:55.888 [INFO][5769] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:55.888 [INFO][5769] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:56.051 [WARNING][5769] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:56.051 [INFO][5769] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" HandleID="k8s-pod-network.0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Workload="localhost-k8s-coredns--7d764666f9--m5vbx-eth0" Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:56.074 [INFO][5769] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:56.107152 containerd[1454]: 2026-04-14 13:33:56.082 [INFO][5759] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9" Apr 14 13:33:56.173861 containerd[1454]: time="2026-04-14T13:33:56.163557703Z" level=info msg="TearDown network for sandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\" successfully" Apr 14 13:33:56.239685 containerd[1454]: time="2026-04-14T13:33:56.239467317Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 14 13:33:56.239685 containerd[1454]: time="2026-04-14T13:33:56.239667705Z" level=info msg="RemovePodSandbox \"0040702ac7bd17ef510d520f6e57f733f732be8b85f99b1cd0475470625a8cf9\" returns successfully" Apr 14 13:33:56.243153 containerd[1454]: time="2026-04-14T13:33:56.242787247Z" level=info msg="StopPodSandbox for \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\"" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.647 [WARNING][5790] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" WorkloadEndpoint="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.662 [INFO][5790] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.662 [INFO][5790] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" iface="eth0" netns="" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.662 [INFO][5790] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.662 [INFO][5790] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.807 [INFO][5799] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.842 [INFO][5799] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.842 [INFO][5799] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.923 [WARNING][5799] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.923 [INFO][5799] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.976 [INFO][5799] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:56.992593 containerd[1454]: 2026-04-14 13:33:56.982 [INFO][5790] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:33:56.995292 containerd[1454]: time="2026-04-14T13:33:56.992768389Z" level=info msg="TearDown network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\" successfully" Apr 14 13:33:56.995292 containerd[1454]: time="2026-04-14T13:33:56.992805614Z" level=info msg="StopPodSandbox for \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\" returns successfully" Apr 14 13:33:56.996920 containerd[1454]: time="2026-04-14T13:33:56.996862645Z" level=info msg="RemovePodSandbox for \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\"" Apr 14 13:33:56.997036 containerd[1454]: time="2026-04-14T13:33:56.996932118Z" level=info msg="Forcibly stopping sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\"" Apr 14 13:33:57.126304 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2820349415.mount: Deactivated successfully. Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.359 [WARNING][5817] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" WorkloadEndpoint="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.361 [INFO][5817] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.361 [INFO][5817] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" iface="eth0" netns="" Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.361 [INFO][5817] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.361 [INFO][5817] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.426 [INFO][5829] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.426 [INFO][5829] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.426 [INFO][5829] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.500 [WARNING][5829] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.501 [INFO][5829] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" HandleID="k8s-pod-network.41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Workload="localhost-k8s-whisker--55d9ccd5cb--vc6bk-eth0" Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.606 [INFO][5829] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:57.618693 containerd[1454]: 2026-04-14 13:33:57.612 [INFO][5817] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78" Apr 14 13:33:57.619976 containerd[1454]: time="2026-04-14T13:33:57.618872755Z" level=info msg="TearDown network for sandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\" successfully" Apr 14 13:33:57.627900 containerd[1454]: time="2026-04-14T13:33:57.627780052Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 14 13:33:57.628650 containerd[1454]: time="2026-04-14T13:33:57.628026317Z" level=info msg="RemovePodSandbox \"41b15e963ae5120443dcc23e17bbfbfe940fc706ec4acd789c9000d203aa4b78\" returns successfully" Apr 14 13:33:57.634446 containerd[1454]: time="2026-04-14T13:33:57.634337341Z" level=info msg="StopPodSandbox for \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\"" Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.132 [WARNING][5846] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--9f7667bb8--268pt-eth0", GenerateName:"goldmane-9f7667bb8-", Namespace:"calico-system", SelfLink:"", UID:"10f4fd58-e76c-4b5c-a2bd-c998eae079c3", ResourceVersion:"1168", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"9f7667bb8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432", Pod:"goldmane-9f7667bb8-268pt", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali2e05b9f506b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.132 [INFO][5846] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.133 [INFO][5846] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" iface="eth0" netns="" Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.133 [INFO][5846] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.133 [INFO][5846] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.285 [INFO][5854] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.287 [INFO][5854] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.287 [INFO][5854] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.388 [WARNING][5854] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.389 [INFO][5854] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.440 [INFO][5854] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:58.467672 containerd[1454]: 2026-04-14 13:33:58.456 [INFO][5846] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:58.467672 containerd[1454]: time="2026-04-14T13:33:58.467385230Z" level=info msg="TearDown network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\" successfully" Apr 14 13:33:58.467672 containerd[1454]: time="2026-04-14T13:33:58.467536521Z" level=info msg="StopPodSandbox for \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\" returns successfully" Apr 14 13:33:58.469690 containerd[1454]: time="2026-04-14T13:33:58.468977590Z" level=info msg="RemovePodSandbox for \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\"" Apr 14 13:33:58.469690 containerd[1454]: time="2026-04-14T13:33:58.469132172Z" level=info msg="Forcibly stopping sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\"" Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.763 [WARNING][5870] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--9f7667bb8--268pt-eth0", GenerateName:"goldmane-9f7667bb8-", Namespace:"calico-system", SelfLink:"", UID:"10f4fd58-e76c-4b5c-a2bd-c998eae079c3", ResourceVersion:"1168", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"9f7667bb8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432", Pod:"goldmane-9f7667bb8-268pt", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali2e05b9f506b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.766 [INFO][5870] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.766 [INFO][5870] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" iface="eth0" netns="" Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.766 [INFO][5870] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.766 [INFO][5870] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.834 [INFO][5878] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.835 [INFO][5878] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.835 [INFO][5878] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.954 [WARNING][5878] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.956 [INFO][5878] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" HandleID="k8s-pod-network.b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Workload="localhost-k8s-goldmane--9f7667bb8--268pt-eth0" Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.996 [INFO][5878] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:59.001873 containerd[1454]: 2026-04-14 13:33:58.998 [INFO][5870] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71" Apr 14 13:33:59.003899 containerd[1454]: time="2026-04-14T13:33:59.002373774Z" level=info msg="TearDown network for sandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\" successfully" Apr 14 13:33:59.022919 containerd[1454]: time="2026-04-14T13:33:59.022859905Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 14 13:33:59.022919 containerd[1454]: time="2026-04-14T13:33:59.022935553Z" level=info msg="RemovePodSandbox \"b3ffdeb5089d6de90c27a8bc19db66ecafc8066cb026fad70a00bcf878611a71\" returns successfully" Apr 14 13:33:59.024574 containerd[1454]: time="2026-04-14T13:33:59.024537126Z" level=info msg="StopPodSandbox for \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\"" Apr 14 13:33:59.074840 containerd[1454]: time="2026-04-14T13:33:59.072001998Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:59.079831 containerd[1454]: time="2026-04-14T13:33:59.079705446Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.4: active requests=0, bytes read=55623386" Apr 14 13:33:59.137813 containerd[1454]: time="2026-04-14T13:33:59.137764636Z" level=info msg="ImageCreate event name:\"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:59.139520 containerd[1454]: time="2026-04-14T13:33:59.139473162Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:59.146808 containerd[1454]: time="2026-04-14T13:33:59.140047761Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" with image id \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\", size \"55623232\" in 11.036058959s" Apr 14 13:33:59.146808 containerd[1454]: time="2026-04-14T13:33:59.140081006Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" returns image reference \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\"" Apr 14 13:33:59.146808 containerd[1454]: time="2026-04-14T13:33:59.146404822Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Apr 14 13:33:59.158334 containerd[1454]: time="2026-04-14T13:33:59.156700401Z" level=info msg="CreateContainer within sandbox \"0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Apr 14 13:33:59.232566 containerd[1454]: time="2026-04-14T13:33:59.232483662Z" level=info msg="CreateContainer within sandbox \"0ddcd30f3227f82102861a1fc914cc40d1d2191c2101b27c25c4d30bf82e9432\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"61451a0b2cd368df6af15af3f655f8e006e0c869516c236b968e103c03bd0a83\"" Apr 14 13:33:59.237586 containerd[1454]: time="2026-04-14T13:33:59.237484097Z" level=info msg="StartContainer for \"61451a0b2cd368df6af15af3f655f8e006e0c869516c236b968e103c03bd0a83\"" Apr 14 13:33:59.393621 systemd[1]: Started cri-containerd-61451a0b2cd368df6af15af3f655f8e006e0c869516c236b968e103c03bd0a83.scope - libcontainer container 61451a0b2cd368df6af15af3f655f8e006e0c869516c236b968e103c03bd0a83. Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.356 [WARNING][5900] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0", GenerateName:"calico-kube-controllers-5bb46d7bb5-", Namespace:"calico-system", SelfLink:"", UID:"2bca4db2-bc4e-488c-88eb-36317594e234", ResourceVersion:"1318", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bb46d7bb5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e", Pod:"calico-kube-controllers-5bb46d7bb5-xwkl7", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib713ab3c088", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.357 [INFO][5900] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.357 [INFO][5900] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" iface="eth0" netns="" Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.357 [INFO][5900] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.357 [INFO][5900] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.405 [INFO][5927] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.425 [INFO][5927] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.426 [INFO][5927] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.458 [WARNING][5927] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.458 [INFO][5927] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.489 [INFO][5927] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:33:59.497004 containerd[1454]: 2026-04-14 13:33:59.491 [INFO][5900] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:33:59.497004 containerd[1454]: time="2026-04-14T13:33:59.496797801Z" level=info msg="TearDown network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\" successfully" Apr 14 13:33:59.497004 containerd[1454]: time="2026-04-14T13:33:59.496830467Z" level=info msg="StopPodSandbox for \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\" returns successfully" Apr 14 13:33:59.502847 containerd[1454]: time="2026-04-14T13:33:59.502751509Z" level=info msg="RemovePodSandbox for \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\"" Apr 14 13:33:59.505530 containerd[1454]: time="2026-04-14T13:33:59.505202637Z" level=info msg="Forcibly stopping sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\"" Apr 14 13:33:59.534124 containerd[1454]: time="2026-04-14T13:33:59.533879828Z" level=info msg="StartContainer for \"61451a0b2cd368df6af15af3f655f8e006e0c869516c236b968e103c03bd0a83\" returns successfully" Apr 14 13:33:59.660715 systemd[1]: Started sshd@13-10.0.0.8:22-10.0.0.1:39492.service - OpenSSH per-connection server daemon (10.0.0.1:39492). Apr 14 13:33:59.723391 containerd[1454]: time="2026-04-14T13:33:59.723294736Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=77" Apr 14 13:33:59.723547 containerd[1454]: time="2026-04-14T13:33:59.723411152Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 14 13:33:59.765466 containerd[1454]: time="2026-04-14T13:33:59.765069766Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"49971841\" in 618.591169ms" Apr 14 13:33:59.765838 containerd[1454]: time="2026-04-14T13:33:59.765715724Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\"" Apr 14 13:33:59.777449 sshd[5970]: Accepted publickey for core from 10.0.0.1 port 39492 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:33:59.781878 sshd[5970]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:33:59.793082 systemd-logind[1442]: New session 14 of user core. Apr 14 13:33:59.796674 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 14 13:33:59.802589 containerd[1454]: time="2026-04-14T13:33:59.802490499Z" level=info msg="CreateContainer within sandbox \"0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 14 13:33:59.904426 containerd[1454]: time="2026-04-14T13:33:59.903850962Z" level=info msg="CreateContainer within sandbox \"0822b5cfd390e64214671cffdc6a7abbdec893d98b7f20f5ab0b32d800e4945f\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"75af7944f04bbba0259f4417014b1624be5b1db3fdf6a8bf0240540c230e2a6c\"" Apr 14 13:33:59.920200 containerd[1454]: time="2026-04-14T13:33:59.918440491Z" level=info msg="StartContainer for \"75af7944f04bbba0259f4417014b1624be5b1db3fdf6a8bf0240540c230e2a6c\"" Apr 14 13:34:00.146998 systemd[1]: Started cri-containerd-75af7944f04bbba0259f4417014b1624be5b1db3fdf6a8bf0240540c230e2a6c.scope - libcontainer container 75af7944f04bbba0259f4417014b1624be5b1db3fdf6a8bf0240540c230e2a6c. Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:33:59.876 [WARNING][5953] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0", GenerateName:"calico-kube-controllers-5bb46d7bb5-", Namespace:"calico-system", SelfLink:"", UID:"2bca4db2-bc4e-488c-88eb-36317594e234", ResourceVersion:"1318", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bb46d7bb5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5d466a5b63580a66af0ca1812a68354f0f47be89db64775ea15c6658c769c54e", Pod:"calico-kube-controllers-5bb46d7bb5-xwkl7", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib713ab3c088", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:33:59.881 [INFO][5953] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:33:59.881 [INFO][5953] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" iface="eth0" netns="" Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:33:59.881 [INFO][5953] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:33:59.881 [INFO][5953] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:33:59.944 [INFO][5978] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:33:59.945 [INFO][5978] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:33:59.947 [INFO][5978] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:34:00.203 [WARNING][5978] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:34:00.203 [INFO][5978] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" HandleID="k8s-pod-network.6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Workload="localhost-k8s-calico--kube--controllers--5bb46d7bb5--xwkl7-eth0" Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:34:00.281 [INFO][5978] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:34:00.362485 containerd[1454]: 2026-04-14 13:34:00.291 [INFO][5953] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970" Apr 14 13:34:00.373954 containerd[1454]: time="2026-04-14T13:34:00.366572349Z" level=info msg="TearDown network for sandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\" successfully" Apr 14 13:34:00.430300 containerd[1454]: time="2026-04-14T13:34:00.428403184Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 14 13:34:00.430300 containerd[1454]: time="2026-04-14T13:34:00.428515611Z" level=info msg="RemovePodSandbox \"6f3e55eace7e7684d316d23810b7263867eaa7e5faa45ae879bdd8713db17970\" returns successfully" Apr 14 13:34:00.431488 containerd[1454]: time="2026-04-14T13:34:00.431442476Z" level=info msg="StopPodSandbox for \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\"" Apr 14 13:34:00.753700 kubelet[2506]: I0414 13:34:00.746131 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/goldmane-9f7667bb8-268pt" podStartSLOduration=74.050159874 podStartE2EDuration="1m51.746115128s" podCreationTimestamp="2026-04-14 13:32:09 +0000 UTC" firstStartedPulling="2026-04-14 13:33:21.446807366 +0000 UTC m=+91.192813754" lastFinishedPulling="2026-04-14 13:33:59.142762628 +0000 UTC m=+128.888769008" observedRunningTime="2026-04-14 13:34:00.740189923 +0000 UTC m=+130.486196307" watchObservedRunningTime="2026-04-14 13:34:00.746115128 +0000 UTC m=+130.492121529" Apr 14 13:34:01.270146 containerd[1454]: time="2026-04-14T13:34:01.269766593Z" level=info msg="StartContainer for \"75af7944f04bbba0259f4417014b1624be5b1db3fdf6a8bf0240540c230e2a6c\" returns successfully" Apr 14 13:34:01.280790 systemd[1]: run-containerd-runc-k8s.io-d4b71d3668f315326a69e2c511d2ce9b7e8f9b7128b92db6c73f0de3adb515f5-runc.8ceaFE.mount: Deactivated successfully. Apr 14 13:34:01.551871 kubelet[2506]: I0414 13:34:01.550843 2506 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-apiserver-56fd9bc5cd-62gjx" podStartSLOduration=80.845189456 podStartE2EDuration="1m53.550706725s" podCreationTimestamp="2026-04-14 13:32:08 +0000 UTC" firstStartedPulling="2026-04-14 13:33:27.072487451 +0000 UTC m=+96.818493831" lastFinishedPulling="2026-04-14 13:33:59.778004719 +0000 UTC m=+129.524011100" observedRunningTime="2026-04-14 13:34:01.536358775 +0000 UTC m=+131.282365162" watchObservedRunningTime="2026-04-14 13:34:01.550706725 +0000 UTC m=+131.296713105" Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.463 [WARNING][6036] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0", GenerateName:"calico-apiserver-56fd9bc5cd-", Namespace:"calico-system", SelfLink:"", UID:"ada6aff6-b823-4c1b-8ba1-78289ce2a978", ResourceVersion:"1264", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56fd9bc5cd", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0", Pod:"calico-apiserver-56fd9bc5cd-hqzvr", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calidbd249e75f3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.468 [INFO][6036] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.468 [INFO][6036] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" iface="eth0" netns="" Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.468 [INFO][6036] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.468 [INFO][6036] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.798 [INFO][6087] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.842 [INFO][6087] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.843 [INFO][6087] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.888 [WARNING][6087] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.890 [INFO][6087] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.912 [INFO][6087] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:34:01.926244 containerd[1454]: 2026-04-14 13:34:01.918 [INFO][6036] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:34:01.927853 containerd[1454]: time="2026-04-14T13:34:01.926332207Z" level=info msg="TearDown network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\" successfully" Apr 14 13:34:01.927853 containerd[1454]: time="2026-04-14T13:34:01.926353890Z" level=info msg="StopPodSandbox for \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\" returns successfully" Apr 14 13:34:01.927853 containerd[1454]: time="2026-04-14T13:34:01.927536441Z" level=info msg="RemovePodSandbox for \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\"" Apr 14 13:34:01.927853 containerd[1454]: time="2026-04-14T13:34:01.927560861Z" level=info msg="Forcibly stopping sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\"" Apr 14 13:34:01.942839 sshd[5970]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:02.002517 systemd[1]: run-containerd-runc-k8s.io-61451a0b2cd368df6af15af3f655f8e006e0c869516c236b968e103c03bd0a83-runc.zzwX3I.mount: Deactivated successfully. Apr 14 13:34:02.048524 systemd[1]: sshd@13-10.0.0.8:22-10.0.0.1:39492.service: Deactivated successfully. Apr 14 13:34:02.073952 systemd[1]: session-14.scope: Deactivated successfully. Apr 14 13:34:02.074861 systemd[1]: session-14.scope: Consumed 1.137s CPU time. Apr 14 13:34:02.075801 systemd-logind[1442]: Session 14 logged out. Waiting for processes to exit. Apr 14 13:34:02.085382 systemd-logind[1442]: Removed session 14. Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.571 [WARNING][6118] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0", GenerateName:"calico-apiserver-56fd9bc5cd-", Namespace:"calico-system", SelfLink:"", UID:"ada6aff6-b823-4c1b-8ba1-78289ce2a978", ResourceVersion:"1264", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56fd9bc5cd", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6ff69efdc8563f88c44ccbdbf6cb278fb73f8117201db7953ba51c195d8e9de0", Pod:"calico-apiserver-56fd9bc5cd-hqzvr", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calidbd249e75f3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.572 [INFO][6118] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.572 [INFO][6118] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" iface="eth0" netns="" Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.572 [INFO][6118] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.572 [INFO][6118] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.865 [INFO][6141] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.884 [INFO][6141] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.884 [INFO][6141] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.976 [WARNING][6141] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.976 [INFO][6141] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" HandleID="k8s-pod-network.5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Workload="localhost-k8s-calico--apiserver--56fd9bc5cd--hqzvr-eth0" Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.985 [INFO][6141] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:34:02.996996 containerd[1454]: 2026-04-14 13:34:02.988 [INFO][6118] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f" Apr 14 13:34:02.996996 containerd[1454]: time="2026-04-14T13:34:02.995709017Z" level=info msg="TearDown network for sandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\" successfully" Apr 14 13:34:03.000767 containerd[1454]: time="2026-04-14T13:34:03.000422237Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 14 13:34:03.000767 containerd[1454]: time="2026-04-14T13:34:03.000505924Z" level=info msg="RemovePodSandbox \"5d03e3a8f427417a092e98d4705fb81633845acb455b17b53ff16b922d556e4f\" returns successfully" Apr 14 13:34:03.002693 containerd[1454]: time="2026-04-14T13:34:03.002647711Z" level=info msg="StopPodSandbox for \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\"" Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.196 [WARNING][6186] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--km2ks-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"9395319e-7ca7-4248-8723-a075a8a69da5", ResourceVersion:"1235", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 31, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff", Pod:"coredns-7d764666f9-km2ks", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1510ebdd2f6", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.198 [INFO][6186] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.198 [INFO][6186] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" iface="eth0" netns="" Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.198 [INFO][6186] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.198 [INFO][6186] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.247 [INFO][6199] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.248 [INFO][6199] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.248 [INFO][6199] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.269 [WARNING][6199] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.269 [INFO][6199] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.280 [INFO][6199] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:34:03.289597 containerd[1454]: 2026-04-14 13:34:03.284 [INFO][6186] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:34:03.289597 containerd[1454]: time="2026-04-14T13:34:03.287843408Z" level=info msg="TearDown network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\" successfully" Apr 14 13:34:03.289597 containerd[1454]: time="2026-04-14T13:34:03.287920417Z" level=info msg="StopPodSandbox for \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\" returns successfully" Apr 14 13:34:03.297825 containerd[1454]: time="2026-04-14T13:34:03.297694680Z" level=info msg="RemovePodSandbox for \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\"" Apr 14 13:34:03.297825 containerd[1454]: time="2026-04-14T13:34:03.297800379Z" level=info msg="Forcibly stopping sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\"" Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:03.756 [WARNING][6222] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7d764666f9--km2ks-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"9395319e-7ca7-4248-8723-a075a8a69da5", ResourceVersion:"1235", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 31, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e877a6d0c5e9da44d831a46cfa2eaa98de35ce85c74fba7cd9f3982e09b4fbff", Pod:"coredns-7d764666f9-km2ks", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1510ebdd2f6", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:03.758 [INFO][6222] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:03.758 [INFO][6222] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" iface="eth0" netns="" Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:03.758 [INFO][6222] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:03.758 [INFO][6222] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:03.941 [INFO][6230] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:03.955 [INFO][6230] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:03.957 [INFO][6230] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:04.158 [WARNING][6230] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:04.168 [INFO][6230] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" HandleID="k8s-pod-network.47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Workload="localhost-k8s-coredns--7d764666f9--km2ks-eth0" Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:04.275 [INFO][6230] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:34:04.300952 containerd[1454]: 2026-04-14 13:34:04.281 [INFO][6222] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c" Apr 14 13:34:04.302461 containerd[1454]: time="2026-04-14T13:34:04.301162298Z" level=info msg="TearDown network for sandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\" successfully" Apr 14 13:34:04.377633 containerd[1454]: time="2026-04-14T13:34:04.376520630Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 14 13:34:04.377633 containerd[1454]: time="2026-04-14T13:34:04.376749283Z" level=info msg="RemovePodSandbox \"47c2657c6d7e09be71596153a7453121af18efd7c3517910ffb1d8ce124f365c\" returns successfully" Apr 14 13:34:04.379186 containerd[1454]: time="2026-04-14T13:34:04.378473980Z" level=info msg="StopPodSandbox for \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\"" Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:04.823 [WARNING][6248] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--5xvdj-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914", ResourceVersion:"1337", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"589b8b8d94", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041", Pod:"csi-node-driver-5xvdj", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali648a174a2cb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:04.823 [INFO][6248] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:04.823 [INFO][6248] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" iface="eth0" netns="" Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:04.823 [INFO][6248] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:04.823 [INFO][6248] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:04.978 [INFO][6257] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:04.979 [INFO][6257] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:04.979 [INFO][6257] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:05.029 [WARNING][6257] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:05.030 [INFO][6257] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:05.075 [INFO][6257] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:34:05.091029 containerd[1454]: 2026-04-14 13:34:05.079 [INFO][6248] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:34:05.099102 containerd[1454]: time="2026-04-14T13:34:05.096753003Z" level=info msg="TearDown network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\" successfully" Apr 14 13:34:05.099102 containerd[1454]: time="2026-04-14T13:34:05.096891950Z" level=info msg="StopPodSandbox for \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\" returns successfully" Apr 14 13:34:05.102779 containerd[1454]: time="2026-04-14T13:34:05.102307561Z" level=info msg="RemovePodSandbox for \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\"" Apr 14 13:34:05.102779 containerd[1454]: time="2026-04-14T13:34:05.102395372Z" level=info msg="Forcibly stopping sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\"" Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.433 [WARNING][6275] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--5xvdj-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b5b0d33b-1b54-43cc-a4d3-cd788e8eb914", ResourceVersion:"1337", Generation:0, CreationTimestamp:time.Date(2026, time.April, 14, 13, 32, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"589b8b8d94", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"542efbf57cce2302f2732e10bd443ec2b09d57ee72640ad867bdec35e739d041", Pod:"csi-node-driver-5xvdj", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali648a174a2cb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.436 [INFO][6275] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.436 [INFO][6275] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" iface="eth0" netns="" Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.436 [INFO][6275] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.436 [INFO][6275] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.636 [INFO][6284] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.636 [INFO][6284] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.636 [INFO][6284] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.707 [WARNING][6284] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.708 [INFO][6284] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" HandleID="k8s-pod-network.c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Workload="localhost-k8s-csi--node--driver--5xvdj-eth0" Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.837 [INFO][6284] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 14 13:34:05.847848 containerd[1454]: 2026-04-14 13:34:05.842 [INFO][6275] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97" Apr 14 13:34:05.849667 containerd[1454]: time="2026-04-14T13:34:05.848100158Z" level=info msg="TearDown network for sandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\" successfully" Apr 14 13:34:05.864181 containerd[1454]: time="2026-04-14T13:34:05.863381339Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 14 13:34:05.864181 containerd[1454]: time="2026-04-14T13:34:05.863874873Z" level=info msg="RemovePodSandbox \"c465ff55a5e061b7d87945fd4246b71b00ab1f0ed25fbe5d790b22036baacc97\" returns successfully" Apr 14 13:34:06.985071 systemd[1]: Started sshd@14-10.0.0.8:22-10.0.0.1:39500.service - OpenSSH per-connection server daemon (10.0.0.1:39500). Apr 14 13:34:07.201173 sshd[6294]: Accepted publickey for core from 10.0.0.1 port 39500 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:07.204035 sshd[6294]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:07.234840 systemd-logind[1442]: New session 15 of user core. Apr 14 13:34:07.248020 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 14 13:34:08.043748 sshd[6294]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:08.067170 systemd[1]: sshd@14-10.0.0.8:22-10.0.0.1:39500.service: Deactivated successfully. Apr 14 13:34:08.090530 systemd[1]: session-15.scope: Deactivated successfully. Apr 14 13:34:08.107705 systemd-logind[1442]: Session 15 logged out. Waiting for processes to exit. Apr 14 13:34:08.153882 systemd-logind[1442]: Removed session 15. Apr 14 13:34:13.090554 systemd[1]: Started sshd@15-10.0.0.8:22-10.0.0.1:38554.service - OpenSSH per-connection server daemon (10.0.0.1:38554). Apr 14 13:34:13.283934 sshd[6309]: Accepted publickey for core from 10.0.0.1 port 38554 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:13.289607 sshd[6309]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:13.380827 systemd-logind[1442]: New session 16 of user core. Apr 14 13:34:13.387922 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 14 13:34:14.572558 sshd[6309]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:14.578022 systemd[1]: sshd@15-10.0.0.8:22-10.0.0.1:38554.service: Deactivated successfully. Apr 14 13:34:14.580433 systemd[1]: session-16.scope: Deactivated successfully. Apr 14 13:34:14.588787 systemd-logind[1442]: Session 16 logged out. Waiting for processes to exit. Apr 14 13:34:14.590447 systemd-logind[1442]: Removed session 16. Apr 14 13:34:16.514352 kubelet[2506]: E0414 13:34:16.504148 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:34:19.673782 systemd[1]: Started sshd@16-10.0.0.8:22-10.0.0.1:39694.service - OpenSSH per-connection server daemon (10.0.0.1:39694). Apr 14 13:34:19.954859 sshd[6345]: Accepted publickey for core from 10.0.0.1 port 39694 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:19.982767 sshd[6345]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:20.064018 systemd-logind[1442]: New session 17 of user core. Apr 14 13:34:20.068142 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 14 13:34:21.486317 kubelet[2506]: E0414 13:34:21.483307 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:34:22.163872 sshd[6345]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:22.188584 systemd[1]: sshd@16-10.0.0.8:22-10.0.0.1:39694.service: Deactivated successfully. Apr 14 13:34:22.210826 systemd[1]: session-17.scope: Deactivated successfully. Apr 14 13:34:22.234212 systemd[1]: session-17.scope: Consumed 1.361s CPU time. Apr 14 13:34:22.240828 systemd-logind[1442]: Session 17 logged out. Waiting for processes to exit. Apr 14 13:34:22.244138 systemd-logind[1442]: Removed session 17. Apr 14 13:34:27.181370 systemd[1]: Started sshd@17-10.0.0.8:22-10.0.0.1:39706.service - OpenSSH per-connection server daemon (10.0.0.1:39706). Apr 14 13:34:27.241265 sshd[6391]: Accepted publickey for core from 10.0.0.1 port 39706 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:27.242709 sshd[6391]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:27.254682 systemd-logind[1442]: New session 18 of user core. Apr 14 13:34:27.270324 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 14 13:34:27.864048 sshd[6391]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:27.867703 systemd[1]: sshd@17-10.0.0.8:22-10.0.0.1:39706.service: Deactivated successfully. Apr 14 13:34:27.872568 systemd[1]: session-18.scope: Deactivated successfully. Apr 14 13:34:27.876731 systemd-logind[1442]: Session 18 logged out. Waiting for processes to exit. Apr 14 13:34:27.881451 systemd-logind[1442]: Removed session 18. Apr 14 13:34:29.489710 kubelet[2506]: E0414 13:34:29.489171 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:34:31.492639 kubelet[2506]: E0414 13:34:31.492510 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:34:32.894064 systemd[1]: Started sshd@18-10.0.0.8:22-10.0.0.1:39566.service - OpenSSH per-connection server daemon (10.0.0.1:39566). Apr 14 13:34:33.004072 sshd[6439]: Accepted publickey for core from 10.0.0.1 port 39566 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:33.008017 sshd[6439]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:33.018782 systemd-logind[1442]: New session 19 of user core. Apr 14 13:34:33.034183 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 14 13:34:34.192841 sshd[6439]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:34.219395 systemd[1]: sshd@18-10.0.0.8:22-10.0.0.1:39566.service: Deactivated successfully. Apr 14 13:34:34.229383 systemd[1]: session-19.scope: Deactivated successfully. Apr 14 13:34:34.232491 systemd-logind[1442]: Session 19 logged out. Waiting for processes to exit. Apr 14 13:34:34.233727 systemd-logind[1442]: Removed session 19. Apr 14 13:34:35.494321 kubelet[2506]: E0414 13:34:35.493099 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:34:39.310282 systemd[1]: Started sshd@19-10.0.0.8:22-10.0.0.1:39572.service - OpenSSH per-connection server daemon (10.0.0.1:39572). Apr 14 13:34:39.503944 sshd[6496]: Accepted publickey for core from 10.0.0.1 port 39572 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:39.511027 sshd[6496]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:39.561622 systemd-logind[1442]: New session 20 of user core. Apr 14 13:34:39.576768 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 14 13:34:40.894057 sshd[6496]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:40.948966 systemd[1]: sshd@19-10.0.0.8:22-10.0.0.1:39572.service: Deactivated successfully. Apr 14 13:34:40.954498 systemd[1]: session-20.scope: Deactivated successfully. Apr 14 13:34:40.962178 systemd-logind[1442]: Session 20 logged out. Waiting for processes to exit. Apr 14 13:34:40.964204 systemd-logind[1442]: Removed session 20. Apr 14 13:34:42.500491 kubelet[2506]: E0414 13:34:42.500063 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:34:45.960035 systemd[1]: Started sshd@20-10.0.0.8:22-10.0.0.1:54578.service - OpenSSH per-connection server daemon (10.0.0.1:54578). Apr 14 13:34:46.140698 sshd[6515]: Accepted publickey for core from 10.0.0.1 port 54578 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:46.145806 sshd[6515]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:46.162293 systemd-logind[1442]: New session 21 of user core. Apr 14 13:34:46.201525 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 14 13:34:47.392901 sshd[6515]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:47.439112 systemd[1]: sshd@20-10.0.0.8:22-10.0.0.1:54578.service: Deactivated successfully. Apr 14 13:34:47.444771 systemd[1]: session-21.scope: Deactivated successfully. Apr 14 13:34:47.453679 systemd-logind[1442]: Session 21 logged out. Waiting for processes to exit. Apr 14 13:34:47.456148 systemd-logind[1442]: Removed session 21. Apr 14 13:34:52.420330 systemd[1]: Started sshd@21-10.0.0.8:22-10.0.0.1:50614.service - OpenSSH per-connection server daemon (10.0.0.1:50614). Apr 14 13:34:52.624016 sshd[6552]: Accepted publickey for core from 10.0.0.1 port 50614 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:52.626175 sshd[6552]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:52.638409 systemd-logind[1442]: New session 22 of user core. Apr 14 13:34:52.659742 systemd[1]: Started session-22.scope - Session 22 of User core. Apr 14 13:34:53.684948 sshd[6552]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:53.736958 systemd[1]: sshd@21-10.0.0.8:22-10.0.0.1:50614.service: Deactivated successfully. Apr 14 13:34:53.755945 systemd[1]: session-22.scope: Deactivated successfully. Apr 14 13:34:53.766650 systemd-logind[1442]: Session 22 logged out. Waiting for processes to exit. Apr 14 13:34:53.825463 systemd[1]: Started sshd@22-10.0.0.8:22-10.0.0.1:50622.service - OpenSSH per-connection server daemon (10.0.0.1:50622). Apr 14 13:34:53.838543 systemd-logind[1442]: Removed session 22. Apr 14 13:34:54.047140 sshd[6568]: Accepted publickey for core from 10.0.0.1 port 50622 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:54.049857 sshd[6568]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:54.067458 systemd-logind[1442]: New session 23 of user core. Apr 14 13:34:54.081049 systemd[1]: Started session-23.scope - Session 23 of User core. Apr 14 13:34:55.373482 sshd[6568]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:55.389746 systemd[1]: sshd@22-10.0.0.8:22-10.0.0.1:50622.service: Deactivated successfully. Apr 14 13:34:55.393264 systemd[1]: session-23.scope: Deactivated successfully. Apr 14 13:34:55.396380 systemd-logind[1442]: Session 23 logged out. Waiting for processes to exit. Apr 14 13:34:55.461687 systemd[1]: Started sshd@23-10.0.0.8:22-10.0.0.1:50632.service - OpenSSH per-connection server daemon (10.0.0.1:50632). Apr 14 13:34:55.474156 systemd-logind[1442]: Removed session 23. Apr 14 13:34:55.536679 sshd[6611]: Accepted publickey for core from 10.0.0.1 port 50632 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:34:55.546625 sshd[6611]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:34:55.575046 systemd-logind[1442]: New session 24 of user core. Apr 14 13:34:55.584158 systemd[1]: Started session-24.scope - Session 24 of User core. Apr 14 13:34:56.575929 sshd[6611]: pam_unix(sshd:session): session closed for user core Apr 14 13:34:56.593185 systemd[1]: sshd@23-10.0.0.8:22-10.0.0.1:50632.service: Deactivated successfully. Apr 14 13:34:56.604611 systemd[1]: session-24.scope: Deactivated successfully. Apr 14 13:34:56.681776 systemd-logind[1442]: Session 24 logged out. Waiting for processes to exit. Apr 14 13:34:56.683168 systemd-logind[1442]: Removed session 24. Apr 14 13:34:57.484178 kubelet[2506]: E0414 13:34:57.483400 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:34:57.858010 update_engine[1445]: I20260414 13:34:57.856067 1445 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Apr 14 13:34:57.858010 update_engine[1445]: I20260414 13:34:57.856937 1445 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Apr 14 13:34:57.866099 update_engine[1445]: I20260414 13:34:57.865532 1445 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Apr 14 13:34:57.867101 update_engine[1445]: I20260414 13:34:57.866856 1445 omaha_request_params.cc:62] Current group set to lts Apr 14 13:34:57.867357 update_engine[1445]: I20260414 13:34:57.867049 1445 update_attempter.cc:499] Already updated boot flags. Skipping. Apr 14 13:34:57.877885 update_engine[1445]: I20260414 13:34:57.867404 1445 update_attempter.cc:643] Scheduling an action processor start. Apr 14 13:34:57.877885 update_engine[1445]: I20260414 13:34:57.867483 1445 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 14 13:34:57.877885 update_engine[1445]: I20260414 13:34:57.867542 1445 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Apr 14 13:34:57.877885 update_engine[1445]: I20260414 13:34:57.867595 1445 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 14 13:34:57.877885 update_engine[1445]: I20260414 13:34:57.867609 1445 omaha_request_action.cc:272] Request: Apr 14 13:34:57.877885 update_engine[1445]: Apr 14 13:34:57.877885 update_engine[1445]: Apr 14 13:34:57.877885 update_engine[1445]: Apr 14 13:34:57.877885 update_engine[1445]: Apr 14 13:34:57.877885 update_engine[1445]: Apr 14 13:34:57.877885 update_engine[1445]: Apr 14 13:34:57.877885 update_engine[1445]: Apr 14 13:34:57.877885 update_engine[1445]: Apr 14 13:34:57.877885 update_engine[1445]: I20260414 13:34:57.867619 1445 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 14 13:34:57.917495 update_engine[1445]: I20260414 13:34:57.911604 1445 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 14 13:34:57.918349 update_engine[1445]: I20260414 13:34:57.918290 1445 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 14 13:34:57.936399 update_engine[1445]: E20260414 13:34:57.931118 1445 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 14 13:34:57.936842 locksmithd[1479]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Apr 14 13:34:57.954105 update_engine[1445]: I20260414 13:34:57.952930 1445 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Apr 14 13:35:01.175994 systemd[1]: run-containerd-runc-k8s.io-d4b71d3668f315326a69e2c511d2ce9b7e8f9b7128b92db6c73f0de3adb515f5-runc.e9aZAe.mount: Deactivated successfully. Apr 14 13:35:01.610207 systemd[1]: Started sshd@24-10.0.0.8:22-10.0.0.1:54474.service - OpenSSH per-connection server daemon (10.0.0.1:54474). Apr 14 13:35:01.842684 sshd[6656]: Accepted publickey for core from 10.0.0.1 port 54474 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:01.845547 sshd[6656]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:01.863874 systemd-logind[1442]: New session 25 of user core. Apr 14 13:35:01.876104 systemd[1]: Started session-25.scope - Session 25 of User core. Apr 14 13:35:03.052784 systemd[1]: run-containerd-runc-k8s.io-61451a0b2cd368df6af15af3f655f8e006e0c869516c236b968e103c03bd0a83-runc.arnzO2.mount: Deactivated successfully. Apr 14 13:35:03.209214 sshd[6656]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:03.236531 systemd[1]: sshd@24-10.0.0.8:22-10.0.0.1:54474.service: Deactivated successfully. Apr 14 13:35:03.258005 systemd[1]: session-25.scope: Deactivated successfully. Apr 14 13:35:03.260818 systemd-logind[1442]: Session 25 logged out. Waiting for processes to exit. Apr 14 13:35:03.266713 systemd-logind[1442]: Removed session 25. Apr 14 13:35:07.807730 update_engine[1445]: I20260414 13:35:07.807467 1445 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 14 13:35:07.809381 update_engine[1445]: I20260414 13:35:07.808585 1445 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 14 13:35:07.816872 update_engine[1445]: I20260414 13:35:07.815811 1445 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 14 13:35:07.823739 update_engine[1445]: E20260414 13:35:07.822547 1445 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 14 13:35:07.823739 update_engine[1445]: I20260414 13:35:07.823532 1445 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Apr 14 13:35:08.252020 systemd[1]: Started sshd@25-10.0.0.8:22-10.0.0.1:54488.service - OpenSSH per-connection server daemon (10.0.0.1:54488). Apr 14 13:35:08.431872 sshd[6716]: Accepted publickey for core from 10.0.0.1 port 54488 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:08.434102 sshd[6716]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:08.462342 systemd-logind[1442]: New session 26 of user core. Apr 14 13:35:08.492135 systemd[1]: Started session-26.scope - Session 26 of User core. Apr 14 13:35:09.886683 sshd[6716]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:09.907884 systemd-logind[1442]: Session 26 logged out. Waiting for processes to exit. Apr 14 13:35:09.907958 systemd[1]: sshd@25-10.0.0.8:22-10.0.0.1:54488.service: Deactivated successfully. Apr 14 13:35:09.911292 systemd[1]: session-26.scope: Deactivated successfully. Apr 14 13:35:09.916161 systemd-logind[1442]: Removed session 26. Apr 14 13:35:14.946286 systemd[1]: Started sshd@26-10.0.0.8:22-10.0.0.1:44752.service - OpenSSH per-connection server daemon (10.0.0.1:44752). Apr 14 13:35:15.143439 sshd[6730]: Accepted publickey for core from 10.0.0.1 port 44752 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:15.150610 sshd[6730]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:15.164482 systemd-logind[1442]: New session 27 of user core. Apr 14 13:35:15.176630 systemd[1]: Started session-27.scope - Session 27 of User core. Apr 14 13:35:16.266458 sshd[6730]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:16.339181 systemd[1]: sshd@26-10.0.0.8:22-10.0.0.1:44752.service: Deactivated successfully. Apr 14 13:35:16.344089 systemd[1]: session-27.scope: Deactivated successfully. Apr 14 13:35:16.350555 systemd-logind[1442]: Session 27 logged out. Waiting for processes to exit. Apr 14 13:35:16.355493 systemd-logind[1442]: Removed session 27. Apr 14 13:35:17.806633 update_engine[1445]: I20260414 13:35:17.806116 1445 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 14 13:35:17.807802 update_engine[1445]: I20260414 13:35:17.807421 1445 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 14 13:35:17.807802 update_engine[1445]: I20260414 13:35:17.807703 1445 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 14 13:35:17.825662 update_engine[1445]: E20260414 13:35:17.825439 1445 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 14 13:35:17.825954 update_engine[1445]: I20260414 13:35:17.825881 1445 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Apr 14 13:35:19.490807 kubelet[2506]: E0414 13:35:19.490755 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:35:21.389433 systemd[1]: Started sshd@27-10.0.0.8:22-10.0.0.1:50020.service - OpenSSH per-connection server daemon (10.0.0.1:50020). Apr 14 13:35:21.451193 sshd[6765]: Accepted publickey for core from 10.0.0.1 port 50020 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:21.455118 sshd[6765]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:21.479809 systemd-logind[1442]: New session 28 of user core. Apr 14 13:35:21.504113 systemd[1]: Started session-28.scope - Session 28 of User core. Apr 14 13:35:22.155729 sshd[6765]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:22.175104 systemd[1]: sshd@27-10.0.0.8:22-10.0.0.1:50020.service: Deactivated successfully. Apr 14 13:35:22.197890 systemd[1]: session-28.scope: Deactivated successfully. Apr 14 13:35:22.199184 systemd-logind[1442]: Session 28 logged out. Waiting for processes to exit. Apr 14 13:35:22.210700 systemd-logind[1442]: Removed session 28. Apr 14 13:35:22.501164 kubelet[2506]: E0414 13:35:22.500878 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:35:27.258150 systemd[1]: Started sshd@28-10.0.0.8:22-10.0.0.1:50026.service - OpenSSH per-connection server daemon (10.0.0.1:50026). Apr 14 13:35:27.443930 sshd[6803]: Accepted publickey for core from 10.0.0.1 port 50026 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:27.463085 sshd[6803]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:27.500387 systemd-logind[1442]: New session 29 of user core. Apr 14 13:35:27.523990 systemd[1]: Started session-29.scope - Session 29 of User core. Apr 14 13:35:27.820084 update_engine[1445]: I20260414 13:35:27.812634 1445 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 14 13:35:27.820084 update_engine[1445]: I20260414 13:35:27.813642 1445 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 14 13:35:27.834772 update_engine[1445]: I20260414 13:35:27.815132 1445 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 14 13:35:27.839966 update_engine[1445]: E20260414 13:35:27.839635 1445 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 14 13:35:27.841568 update_engine[1445]: I20260414 13:35:27.841464 1445 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 14 13:35:27.841713 update_engine[1445]: I20260414 13:35:27.841699 1445 omaha_request_action.cc:617] Omaha request response: Apr 14 13:35:27.848681 update_engine[1445]: E20260414 13:35:27.844892 1445 omaha_request_action.cc:636] Omaha request network transfer failed. Apr 14 13:35:27.849732 update_engine[1445]: I20260414 13:35:27.849565 1445 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Apr 14 13:35:27.849915 update_engine[1445]: I20260414 13:35:27.849898 1445 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 14 13:35:27.849959 update_engine[1445]: I20260414 13:35:27.849949 1445 update_attempter.cc:306] Processing Done. Apr 14 13:35:27.854195 update_engine[1445]: E20260414 13:35:27.853933 1445 update_attempter.cc:619] Update failed. Apr 14 13:35:27.855259 update_engine[1445]: I20260414 13:35:27.854588 1445 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Apr 14 13:35:27.855259 update_engine[1445]: I20260414 13:35:27.854609 1445 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Apr 14 13:35:27.855259 update_engine[1445]: I20260414 13:35:27.854624 1445 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Apr 14 13:35:27.855259 update_engine[1445]: I20260414 13:35:27.854848 1445 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 14 13:35:27.855259 update_engine[1445]: I20260414 13:35:27.854990 1445 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 14 13:35:27.855259 update_engine[1445]: I20260414 13:35:27.854995 1445 omaha_request_action.cc:272] Request: Apr 14 13:35:27.855259 update_engine[1445]: Apr 14 13:35:27.855259 update_engine[1445]: Apr 14 13:35:27.855259 update_engine[1445]: Apr 14 13:35:27.855259 update_engine[1445]: Apr 14 13:35:27.855259 update_engine[1445]: Apr 14 13:35:27.855259 update_engine[1445]: Apr 14 13:35:27.855259 update_engine[1445]: I20260414 13:35:27.855002 1445 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 14 13:35:27.857692 update_engine[1445]: I20260414 13:35:27.855938 1445 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 14 13:35:27.857692 update_engine[1445]: I20260414 13:35:27.856663 1445 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 14 13:35:27.861643 locksmithd[1479]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Apr 14 13:35:27.869416 update_engine[1445]: E20260414 13:35:27.868736 1445 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 14 13:35:27.869416 update_engine[1445]: I20260414 13:35:27.869019 1445 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 14 13:35:27.869416 update_engine[1445]: I20260414 13:35:27.869037 1445 omaha_request_action.cc:617] Omaha request response: Apr 14 13:35:27.869416 update_engine[1445]: I20260414 13:35:27.869052 1445 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 14 13:35:27.869416 update_engine[1445]: I20260414 13:35:27.869058 1445 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 14 13:35:27.869416 update_engine[1445]: I20260414 13:35:27.869063 1445 update_attempter.cc:306] Processing Done. Apr 14 13:35:27.869416 update_engine[1445]: I20260414 13:35:27.869074 1445 update_attempter.cc:310] Error event sent. Apr 14 13:35:27.869416 update_engine[1445]: I20260414 13:35:27.869096 1445 update_check_scheduler.cc:74] Next update check in 42m52s Apr 14 13:35:27.876895 locksmithd[1479]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Apr 14 13:35:28.733737 sshd[6803]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:28.759944 systemd[1]: sshd@28-10.0.0.8:22-10.0.0.1:50026.service: Deactivated successfully. Apr 14 13:35:28.799896 systemd[1]: session-29.scope: Deactivated successfully. Apr 14 13:35:28.873846 systemd-logind[1442]: Session 29 logged out. Waiting for processes to exit. Apr 14 13:35:28.875639 systemd-logind[1442]: Removed session 29. Apr 14 13:35:33.836198 systemd[1]: Started sshd@29-10.0.0.8:22-10.0.0.1:55350.service - OpenSSH per-connection server daemon (10.0.0.1:55350). Apr 14 13:35:34.059660 sshd[6838]: Accepted publickey for core from 10.0.0.1 port 55350 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:34.074725 sshd[6838]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:34.171492 systemd-logind[1442]: New session 30 of user core. Apr 14 13:35:34.182038 systemd[1]: Started session-30.scope - Session 30 of User core. Apr 14 13:35:35.381129 sshd[6838]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:35.405032 systemd-logind[1442]: Session 30 logged out. Waiting for processes to exit. Apr 14 13:35:35.466992 systemd[1]: sshd@29-10.0.0.8:22-10.0.0.1:55350.service: Deactivated successfully. Apr 14 13:35:35.494170 systemd[1]: session-30.scope: Deactivated successfully. Apr 14 13:35:35.518139 systemd-logind[1442]: Removed session 30. Apr 14 13:35:39.506444 kubelet[2506]: E0414 13:35:39.501973 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:35:40.474578 systemd[1]: Started sshd@30-10.0.0.8:22-10.0.0.1:58812.service - OpenSSH per-connection server daemon (10.0.0.1:58812). Apr 14 13:35:40.620677 sshd[6858]: Accepted publickey for core from 10.0.0.1 port 58812 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:40.626030 sshd[6858]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:40.657576 systemd-logind[1442]: New session 31 of user core. Apr 14 13:35:40.672077 systemd[1]: Started session-31.scope - Session 31 of User core. Apr 14 13:35:41.464774 sshd[6858]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:41.480934 systemd[1]: sshd@30-10.0.0.8:22-10.0.0.1:58812.service: Deactivated successfully. Apr 14 13:35:41.502911 systemd[1]: session-31.scope: Deactivated successfully. Apr 14 13:35:41.509439 systemd-logind[1442]: Session 31 logged out. Waiting for processes to exit. Apr 14 13:35:41.510870 systemd-logind[1442]: Removed session 31. Apr 14 13:35:44.585302 kubelet[2506]: E0414 13:35:44.585178 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:35:46.557511 systemd[1]: Started sshd@31-10.0.0.8:22-10.0.0.1:58820.service - OpenSSH per-connection server daemon (10.0.0.1:58820). Apr 14 13:35:46.906564 sshd[6890]: Accepted publickey for core from 10.0.0.1 port 58820 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:46.931097 sshd[6890]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:46.997791 systemd-logind[1442]: New session 32 of user core. Apr 14 13:35:47.074015 systemd[1]: Started session-32.scope - Session 32 of User core. Apr 14 13:35:48.118989 sshd[6890]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:48.142319 systemd[1]: sshd@31-10.0.0.8:22-10.0.0.1:58820.service: Deactivated successfully. Apr 14 13:35:48.165659 systemd[1]: session-32.scope: Deactivated successfully. Apr 14 13:35:48.182122 systemd-logind[1442]: Session 32 logged out. Waiting for processes to exit. Apr 14 13:35:48.204162 systemd-logind[1442]: Removed session 32. Apr 14 13:35:53.160137 systemd[1]: Started sshd@32-10.0.0.8:22-10.0.0.1:45902.service - OpenSSH per-connection server daemon (10.0.0.1:45902). Apr 14 13:35:53.277531 sshd[6925]: Accepted publickey for core from 10.0.0.1 port 45902 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:53.288538 sshd[6925]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:53.305753 systemd-logind[1442]: New session 33 of user core. Apr 14 13:35:53.334655 systemd[1]: Started session-33.scope - Session 33 of User core. Apr 14 13:35:54.482748 kubelet[2506]: E0414 13:35:54.482688 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:35:54.514244 sshd[6925]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:54.529330 systemd[1]: sshd@32-10.0.0.8:22-10.0.0.1:45902.service: Deactivated successfully. Apr 14 13:35:54.577865 systemd[1]: session-33.scope: Deactivated successfully. Apr 14 13:35:54.579560 systemd-logind[1442]: Session 33 logged out. Waiting for processes to exit. Apr 14 13:35:54.605607 systemd[1]: Started sshd@33-10.0.0.8:22-10.0.0.1:45908.service - OpenSSH per-connection server daemon (10.0.0.1:45908). Apr 14 13:35:54.609764 systemd-logind[1442]: Removed session 33. Apr 14 13:35:54.888288 sshd[6961]: Accepted publickey for core from 10.0.0.1 port 45908 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:54.894252 sshd[6961]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:54.992243 systemd-logind[1442]: New session 34 of user core. Apr 14 13:35:55.004576 systemd[1]: Started session-34.scope - Session 34 of User core. Apr 14 13:35:57.048168 sshd[6961]: pam_unix(sshd:session): session closed for user core Apr 14 13:35:57.108063 systemd[1]: sshd@33-10.0.0.8:22-10.0.0.1:45908.service: Deactivated successfully. Apr 14 13:35:57.130343 systemd[1]: session-34.scope: Deactivated successfully. Apr 14 13:35:57.131831 systemd[1]: session-34.scope: Consumed 1.218s CPU time. Apr 14 13:35:57.133764 systemd-logind[1442]: Session 34 logged out. Waiting for processes to exit. Apr 14 13:35:57.153928 systemd[1]: Started sshd@34-10.0.0.8:22-10.0.0.1:45918.service - OpenSSH per-connection server daemon (10.0.0.1:45918). Apr 14 13:35:57.159463 systemd-logind[1442]: Removed session 34. Apr 14 13:35:57.275098 sshd[6976]: Accepted publickey for core from 10.0.0.1 port 45918 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:35:57.279719 sshd[6976]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:35:57.322335 systemd-logind[1442]: New session 35 of user core. Apr 14 13:35:57.341520 systemd[1]: Started session-35.scope - Session 35 of User core. Apr 14 13:35:59.487548 kubelet[2506]: E0414 13:35:59.486700 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:36:01.161507 systemd[1]: run-containerd-runc-k8s.io-d4b71d3668f315326a69e2c511d2ce9b7e8f9b7128b92db6c73f0de3adb515f5-runc.KinE01.mount: Deactivated successfully. Apr 14 13:36:02.088698 systemd[1]: run-containerd-runc-k8s.io-61451a0b2cd368df6af15af3f655f8e006e0c869516c236b968e103c03bd0a83-runc.IgioBG.mount: Deactivated successfully. Apr 14 13:36:02.931971 sshd[6976]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:02.944892 systemd[1]: Started sshd@35-10.0.0.8:22-10.0.0.1:53978.service - OpenSSH per-connection server daemon (10.0.0.1:53978). Apr 14 13:36:02.961876 systemd[1]: sshd@34-10.0.0.8:22-10.0.0.1:45918.service: Deactivated successfully. Apr 14 13:36:02.980395 systemd[1]: session-35.scope: Deactivated successfully. Apr 14 13:36:02.980567 systemd[1]: session-35.scope: Consumed 2.775s CPU time. Apr 14 13:36:02.984738 systemd-logind[1442]: Session 35 logged out. Waiting for processes to exit. Apr 14 13:36:02.986760 systemd-logind[1442]: Removed session 35. Apr 14 13:36:03.045658 sshd[7048]: Accepted publickey for core from 10.0.0.1 port 53978 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:03.048018 sshd[7048]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:03.062301 systemd-logind[1442]: New session 36 of user core. Apr 14 13:36:03.072021 systemd[1]: Started session-36.scope - Session 36 of User core. Apr 14 13:36:06.859815 sshd[7048]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:06.884199 systemd[1]: Started sshd@36-10.0.0.8:22-10.0.0.1:53986.service - OpenSSH per-connection server daemon (10.0.0.1:53986). Apr 14 13:36:06.942953 systemd[1]: sshd@35-10.0.0.8:22-10.0.0.1:53978.service: Deactivated successfully. Apr 14 13:36:06.977106 systemd[1]: session-36.scope: Deactivated successfully. Apr 14 13:36:06.977748 systemd[1]: session-36.scope: Consumed 2.302s CPU time. Apr 14 13:36:07.027771 systemd-logind[1442]: Session 36 logged out. Waiting for processes to exit. Apr 14 13:36:07.034672 systemd-logind[1442]: Removed session 36. Apr 14 13:36:07.084731 sshd[7081]: Accepted publickey for core from 10.0.0.1 port 53986 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:07.102388 sshd[7081]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:07.157634 systemd-logind[1442]: New session 37 of user core. Apr 14 13:36:07.202130 systemd[1]: Started session-37.scope - Session 37 of User core. Apr 14 13:36:08.009337 sshd[7081]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:08.056276 systemd-logind[1442]: Session 37 logged out. Waiting for processes to exit. Apr 14 13:36:08.056537 systemd[1]: sshd@36-10.0.0.8:22-10.0.0.1:53986.service: Deactivated successfully. Apr 14 13:36:08.058722 systemd[1]: session-37.scope: Deactivated successfully. Apr 14 13:36:08.062131 systemd-logind[1442]: Removed session 37. Apr 14 13:36:13.121757 systemd[1]: Started sshd@37-10.0.0.8:22-10.0.0.1:54970.service - OpenSSH per-connection server daemon (10.0.0.1:54970). Apr 14 13:36:13.278763 sshd[7127]: Accepted publickey for core from 10.0.0.1 port 54970 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:13.283077 sshd[7127]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:13.330741 systemd-logind[1442]: New session 38 of user core. Apr 14 13:36:13.368811 systemd[1]: Started session-38.scope - Session 38 of User core. Apr 14 13:36:14.203272 sshd[7127]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:14.243289 systemd[1]: sshd@37-10.0.0.8:22-10.0.0.1:54970.service: Deactivated successfully. Apr 14 13:36:14.250660 systemd[1]: session-38.scope: Deactivated successfully. Apr 14 13:36:14.255032 systemd-logind[1442]: Session 38 logged out. Waiting for processes to exit. Apr 14 13:36:14.261492 systemd-logind[1442]: Removed session 38. Apr 14 13:36:19.305609 systemd[1]: Started sshd@38-10.0.0.8:22-10.0.0.1:54978.service - OpenSSH per-connection server daemon (10.0.0.1:54978). Apr 14 13:36:19.449347 sshd[7165]: Accepted publickey for core from 10.0.0.1 port 54978 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:19.461559 sshd[7165]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:19.489082 systemd-logind[1442]: New session 39 of user core. Apr 14 13:36:19.572596 systemd[1]: Started session-39.scope - Session 39 of User core. Apr 14 13:36:20.969603 sshd[7165]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:20.973690 systemd[1]: sshd@38-10.0.0.8:22-10.0.0.1:54978.service: Deactivated successfully. Apr 14 13:36:20.980197 systemd[1]: session-39.scope: Deactivated successfully. Apr 14 13:36:20.981115 systemd-logind[1442]: Session 39 logged out. Waiting for processes to exit. Apr 14 13:36:20.982310 systemd-logind[1442]: Removed session 39. Apr 14 13:36:25.493516 kubelet[2506]: E0414 13:36:25.493358 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:36:26.057653 systemd[1]: Started sshd@39-10.0.0.8:22-10.0.0.1:56244.service - OpenSSH per-connection server daemon (10.0.0.1:56244). Apr 14 13:36:26.221442 sshd[7205]: Accepted publickey for core from 10.0.0.1 port 56244 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:26.223273 sshd[7205]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:26.308148 systemd-logind[1442]: New session 40 of user core. Apr 14 13:36:26.328573 systemd[1]: Started session-40.scope - Session 40 of User core. Apr 14 13:36:27.548991 sshd[7205]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:27.563879 systemd[1]: sshd@39-10.0.0.8:22-10.0.0.1:56244.service: Deactivated successfully. Apr 14 13:36:27.575923 systemd[1]: session-40.scope: Deactivated successfully. Apr 14 13:36:27.595147 systemd-logind[1442]: Session 40 logged out. Waiting for processes to exit. Apr 14 13:36:27.603964 systemd-logind[1442]: Removed session 40. Apr 14 13:36:30.488038 kubelet[2506]: E0414 13:36:30.487953 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:36:32.606093 systemd[1]: Started sshd@40-10.0.0.8:22-10.0.0.1:48494.service - OpenSSH per-connection server daemon (10.0.0.1:48494). Apr 14 13:36:32.817162 sshd[7243]: Accepted publickey for core from 10.0.0.1 port 48494 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:32.851422 sshd[7243]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:32.957753 systemd-logind[1442]: New session 41 of user core. Apr 14 13:36:32.965166 systemd[1]: Started session-41.scope - Session 41 of User core. Apr 14 13:36:33.584447 sshd[7243]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:33.595822 systemd[1]: sshd@40-10.0.0.8:22-10.0.0.1:48494.service: Deactivated successfully. Apr 14 13:36:33.604568 systemd[1]: session-41.scope: Deactivated successfully. Apr 14 13:36:33.605593 systemd-logind[1442]: Session 41 logged out. Waiting for processes to exit. Apr 14 13:36:33.613951 systemd-logind[1442]: Removed session 41. Apr 14 13:36:38.646066 systemd[1]: Started sshd@41-10.0.0.8:22-10.0.0.1:48508.service - OpenSSH per-connection server daemon (10.0.0.1:48508). Apr 14 13:36:38.826073 sshd[7257]: Accepted publickey for core from 10.0.0.1 port 48508 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:38.834855 sshd[7257]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:38.860713 systemd-logind[1442]: New session 42 of user core. Apr 14 13:36:38.879705 systemd[1]: Started session-42.scope - Session 42 of User core. Apr 14 13:36:39.653942 sshd[7257]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:39.663564 systemd[1]: sshd@41-10.0.0.8:22-10.0.0.1:48508.service: Deactivated successfully. Apr 14 13:36:39.669294 systemd[1]: session-42.scope: Deactivated successfully. Apr 14 13:36:39.679996 systemd-logind[1442]: Session 42 logged out. Waiting for processes to exit. Apr 14 13:36:39.699033 systemd-logind[1442]: Removed session 42. Apr 14 13:36:40.507758 kubelet[2506]: E0414 13:36:40.507715 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:36:44.739083 systemd[1]: Started sshd@42-10.0.0.8:22-10.0.0.1:58408.service - OpenSSH per-connection server daemon (10.0.0.1:58408). Apr 14 13:36:44.964304 sshd[7272]: Accepted publickey for core from 10.0.0.1 port 58408 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:44.972111 sshd[7272]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:45.051859 systemd-logind[1442]: New session 43 of user core. Apr 14 13:36:45.064633 systemd[1]: Started session-43.scope - Session 43 of User core. Apr 14 13:36:46.171636 sshd[7272]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:46.203072 systemd[1]: sshd@42-10.0.0.8:22-10.0.0.1:58408.service: Deactivated successfully. Apr 14 13:36:46.260778 systemd[1]: session-43.scope: Deactivated successfully. Apr 14 13:36:46.276999 systemd-logind[1442]: Session 43 logged out. Waiting for processes to exit. Apr 14 13:36:46.288159 systemd-logind[1442]: Removed session 43. Apr 14 13:36:49.492639 kubelet[2506]: E0414 13:36:49.488984 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:36:51.224374 systemd[1]: Started sshd@43-10.0.0.8:22-10.0.0.1:51748.service - OpenSSH per-connection server daemon (10.0.0.1:51748). Apr 14 13:36:51.276571 sshd[7311]: Accepted publickey for core from 10.0.0.1 port 51748 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:51.278990 sshd[7311]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:51.301616 systemd-logind[1442]: New session 44 of user core. Apr 14 13:36:51.314461 systemd[1]: Started session-44.scope - Session 44 of User core. Apr 14 13:36:52.262993 sshd[7311]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:52.270550 systemd[1]: sshd@43-10.0.0.8:22-10.0.0.1:51748.service: Deactivated successfully. Apr 14 13:36:52.277832 systemd[1]: session-44.scope: Deactivated successfully. Apr 14 13:36:52.278800 systemd-logind[1442]: Session 44 logged out. Waiting for processes to exit. Apr 14 13:36:52.281869 systemd-logind[1442]: Removed session 44. Apr 14 13:36:57.348390 systemd[1]: Started sshd@44-10.0.0.8:22-10.0.0.1:51752.service - OpenSSH per-connection server daemon (10.0.0.1:51752). Apr 14 13:36:57.475854 sshd[7351]: Accepted publickey for core from 10.0.0.1 port 51752 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:36:57.479595 sshd[7351]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:36:57.498160 systemd-logind[1442]: New session 45 of user core. Apr 14 13:36:57.508369 systemd[1]: Started session-45.scope - Session 45 of User core. Apr 14 13:36:58.597556 sshd[7351]: pam_unix(sshd:session): session closed for user core Apr 14 13:36:58.710486 systemd-logind[1442]: Session 45 logged out. Waiting for processes to exit. Apr 14 13:36:58.717212 systemd[1]: sshd@44-10.0.0.8:22-10.0.0.1:51752.service: Deactivated successfully. Apr 14 13:36:58.723713 systemd[1]: session-45.scope: Deactivated successfully. Apr 14 13:36:58.739362 systemd-logind[1442]: Removed session 45. Apr 14 13:37:03.718123 systemd[1]: Started sshd@45-10.0.0.8:22-10.0.0.1:40770.service - OpenSSH per-connection server daemon (10.0.0.1:40770). Apr 14 13:37:03.913729 sshd[7429]: Accepted publickey for core from 10.0.0.1 port 40770 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:03.917189 sshd[7429]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:03.940080 systemd-logind[1442]: New session 46 of user core. Apr 14 13:37:03.953702 systemd[1]: Started session-46.scope - Session 46 of User core. Apr 14 13:37:05.197864 sshd[7429]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:05.237752 systemd-logind[1442]: Session 46 logged out. Waiting for processes to exit. Apr 14 13:37:05.261780 systemd[1]: sshd@45-10.0.0.8:22-10.0.0.1:40770.service: Deactivated successfully. Apr 14 13:37:05.282859 systemd[1]: session-46.scope: Deactivated successfully. Apr 14 13:37:05.292543 systemd-logind[1442]: Removed session 46. Apr 14 13:37:10.289749 systemd[1]: Started sshd@46-10.0.0.8:22-10.0.0.1:57848.service - OpenSSH per-connection server daemon (10.0.0.1:57848). Apr 14 13:37:10.429399 sshd[7443]: Accepted publickey for core from 10.0.0.1 port 57848 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:10.435495 sshd[7443]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:10.488808 systemd-logind[1442]: New session 47 of user core. Apr 14 13:37:10.508646 systemd[1]: Started session-47.scope - Session 47 of User core. Apr 14 13:37:11.442190 sshd[7443]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:11.455527 systemd[1]: sshd@46-10.0.0.8:22-10.0.0.1:57848.service: Deactivated successfully. Apr 14 13:37:11.461524 systemd[1]: session-47.scope: Deactivated successfully. Apr 14 13:37:11.484461 systemd-logind[1442]: Session 47 logged out. Waiting for processes to exit. Apr 14 13:37:11.487770 systemd-logind[1442]: Removed session 47. Apr 14 13:37:14.497773 kubelet[2506]: E0414 13:37:14.497109 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:37:16.492675 systemd[1]: Started sshd@47-10.0.0.8:22-10.0.0.1:57850.service - OpenSSH per-connection server daemon (10.0.0.1:57850). Apr 14 13:37:16.651447 sshd[7473]: Accepted publickey for core from 10.0.0.1 port 57850 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:16.653443 sshd[7473]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:16.794565 systemd-logind[1442]: New session 48 of user core. Apr 14 13:37:16.810423 systemd[1]: Started session-48.scope - Session 48 of User core. Apr 14 13:37:17.493629 sshd[7473]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:17.509026 systemd[1]: sshd@47-10.0.0.8:22-10.0.0.1:57850.service: Deactivated successfully. Apr 14 13:37:17.525654 systemd[1]: session-48.scope: Deactivated successfully. Apr 14 13:37:17.534399 systemd-logind[1442]: Session 48 logged out. Waiting for processes to exit. Apr 14 13:37:17.539169 systemd-logind[1442]: Removed session 48. Apr 14 13:37:18.547576 kubelet[2506]: E0414 13:37:18.547433 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:37:22.535449 systemd[1]: Started sshd@48-10.0.0.8:22-10.0.0.1:39652.service - OpenSSH per-connection server daemon (10.0.0.1:39652). Apr 14 13:37:22.671269 sshd[7494]: Accepted publickey for core from 10.0.0.1 port 39652 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:22.673066 sshd[7494]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:22.688344 systemd-logind[1442]: New session 49 of user core. Apr 14 13:37:22.705717 systemd[1]: Started session-49.scope - Session 49 of User core. Apr 14 13:37:23.742442 sshd[7494]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:23.759622 systemd[1]: sshd@48-10.0.0.8:22-10.0.0.1:39652.service: Deactivated successfully. Apr 14 13:37:23.769035 systemd[1]: session-49.scope: Deactivated successfully. Apr 14 13:37:23.771370 systemd-logind[1442]: Session 49 logged out. Waiting for processes to exit. Apr 14 13:37:23.780079 systemd-logind[1442]: Removed session 49. Apr 14 13:37:27.510389 kubelet[2506]: E0414 13:37:27.510078 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:37:28.784611 systemd[1]: Started sshd@49-10.0.0.8:22-10.0.0.1:39654.service - OpenSSH per-connection server daemon (10.0.0.1:39654). Apr 14 13:37:29.063076 sshd[7533]: Accepted publickey for core from 10.0.0.1 port 39654 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:29.067711 sshd[7533]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:29.109121 systemd-logind[1442]: New session 50 of user core. Apr 14 13:37:29.135627 systemd[1]: Started session-50.scope - Session 50 of User core. Apr 14 13:37:30.169482 sshd[7533]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:30.176134 systemd[1]: sshd@49-10.0.0.8:22-10.0.0.1:39654.service: Deactivated successfully. Apr 14 13:37:30.192209 systemd[1]: session-50.scope: Deactivated successfully. Apr 14 13:37:30.202439 systemd-logind[1442]: Session 50 logged out. Waiting for processes to exit. Apr 14 13:37:30.208625 systemd-logind[1442]: Removed session 50. Apr 14 13:37:35.294313 systemd[1]: Started sshd@50-10.0.0.8:22-10.0.0.1:46890.service - OpenSSH per-connection server daemon (10.0.0.1:46890). Apr 14 13:37:35.487778 kubelet[2506]: E0414 13:37:35.486710 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:37:35.511913 sshd[7568]: Accepted publickey for core from 10.0.0.1 port 46890 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:35.526691 sshd[7568]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:35.579791 systemd-logind[1442]: New session 51 of user core. Apr 14 13:37:35.594861 systemd[1]: Started session-51.scope - Session 51 of User core. Apr 14 13:37:36.807830 sshd[7568]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:36.887823 systemd[1]: sshd@50-10.0.0.8:22-10.0.0.1:46890.service: Deactivated successfully. Apr 14 13:37:36.890353 systemd-logind[1442]: Session 51 logged out. Waiting for processes to exit. Apr 14 13:37:36.898721 systemd[1]: session-51.scope: Deactivated successfully. Apr 14 13:37:36.909894 systemd-logind[1442]: Removed session 51. Apr 14 13:37:41.908924 systemd[1]: Started sshd@51-10.0.0.8:22-10.0.0.1:50638.service - OpenSSH per-connection server daemon (10.0.0.1:50638). Apr 14 13:37:42.121708 sshd[7589]: Accepted publickey for core from 10.0.0.1 port 50638 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:42.125069 sshd[7589]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:42.198748 systemd-logind[1442]: New session 52 of user core. Apr 14 13:37:42.255691 systemd[1]: Started session-52.scope - Session 52 of User core. Apr 14 13:37:43.906325 sshd[7589]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:43.939367 systemd[1]: sshd@51-10.0.0.8:22-10.0.0.1:50638.service: Deactivated successfully. Apr 14 13:37:43.963930 systemd[1]: session-52.scope: Deactivated successfully. Apr 14 13:37:43.965550 systemd[1]: session-52.scope: Consumed 1.024s CPU time. Apr 14 13:37:43.990903 systemd-logind[1442]: Session 52 logged out. Waiting for processes to exit. Apr 14 13:37:44.028667 systemd-logind[1442]: Removed session 52. Apr 14 13:37:46.539243 kubelet[2506]: E0414 13:37:46.539150 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:37:48.964339 systemd[1]: Started sshd@52-10.0.0.8:22-10.0.0.1:50652.service - OpenSSH per-connection server daemon (10.0.0.1:50652). Apr 14 13:37:49.166667 sshd[7658]: Accepted publickey for core from 10.0.0.1 port 50652 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:49.188850 sshd[7658]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:49.251843 systemd-logind[1442]: New session 53 of user core. Apr 14 13:37:49.261317 systemd[1]: Started session-53.scope - Session 53 of User core. Apr 14 13:37:50.811129 sshd[7658]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:50.865383 systemd[1]: sshd@52-10.0.0.8:22-10.0.0.1:50652.service: Deactivated successfully. Apr 14 13:37:50.892400 systemd[1]: session-53.scope: Deactivated successfully. Apr 14 13:37:50.894251 systemd[1]: session-53.scope: Consumed 1.093s CPU time. Apr 14 13:37:50.917026 systemd-logind[1442]: Session 53 logged out. Waiting for processes to exit. Apr 14 13:37:50.924290 systemd-logind[1442]: Removed session 53. Apr 14 13:37:54.510407 kubelet[2506]: E0414 13:37:54.510014 2506 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 14 13:37:55.854106 systemd[1]: Started sshd@53-10.0.0.8:22-10.0.0.1:54416.service - OpenSSH per-connection server daemon (10.0.0.1:54416). Apr 14 13:37:56.209312 sshd[7698]: Accepted publickey for core from 10.0.0.1 port 54416 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:37:56.213196 sshd[7698]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:37:56.238819 systemd-logind[1442]: New session 54 of user core. Apr 14 13:37:56.275275 systemd[1]: Started session-54.scope - Session 54 of User core. Apr 14 13:37:58.445284 sshd[7698]: pam_unix(sshd:session): session closed for user core Apr 14 13:37:58.466610 systemd[1]: sshd@53-10.0.0.8:22-10.0.0.1:54416.service: Deactivated successfully. Apr 14 13:37:58.489604 systemd[1]: session-54.scope: Deactivated successfully. Apr 14 13:37:58.491799 systemd[1]: session-54.scope: Consumed 1.472s CPU time. Apr 14 13:37:58.502002 systemd-logind[1442]: Session 54 logged out. Waiting for processes to exit. Apr 14 13:37:58.506725 systemd-logind[1442]: Removed session 54. Apr 14 13:38:03.549719 systemd[1]: Started sshd@54-10.0.0.8:22-10.0.0.1:47280.service - OpenSSH per-connection server daemon (10.0.0.1:47280). Apr 14 13:38:03.765510 sshd[7782]: Accepted publickey for core from 10.0.0.1 port 47280 ssh2: RSA SHA256:STqg7NDKHqB1pC6cv1a9vkNfz6oKwIzfWFn4Twt++GI Apr 14 13:38:03.766131 sshd[7782]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 14 13:38:03.783008 systemd-logind[1442]: New session 55 of user core. Apr 14 13:38:03.789685 systemd[1]: Started session-55.scope - Session 55 of User core. Apr 14 13:38:05.095934 sshd[7782]: pam_unix(sshd:session): session closed for user core Apr 14 13:38:05.105701 systemd[1]: sshd@54-10.0.0.8:22-10.0.0.1:47280.service: Deactivated successfully. Apr 14 13:38:05.125925 systemd[1]: session-55.scope: Deactivated successfully. Apr 14 13:38:05.132860 systemd-logind[1442]: Session 55 logged out. Waiting for processes to exit. Apr 14 13:38:05.149867 systemd-logind[1442]: Removed session 55.