Oct 28 05:17:40.675614 kernel: Linux version 6.12.54-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Tue Oct 28 03:19:40 -00 2025 Oct 28 05:17:40.675642 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=449db75fd0bf4f00a7b0da93783dc37f82f4a66df937e11c006397de0369495c Oct 28 05:17:40.675652 kernel: BIOS-provided physical RAM map: Oct 28 05:17:40.675667 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Oct 28 05:17:40.675673 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Oct 28 05:17:40.675680 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Oct 28 05:17:40.675688 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Oct 28 05:17:40.675695 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Oct 28 05:17:40.675705 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Oct 28 05:17:40.675711 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Oct 28 05:17:40.675718 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Oct 28 05:17:40.675732 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Oct 28 05:17:40.675739 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Oct 28 05:17:40.675746 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Oct 28 05:17:40.675754 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Oct 28 05:17:40.675762 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Oct 28 05:17:40.675778 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Oct 28 05:17:40.675786 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Oct 28 05:17:40.675793 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Oct 28 05:17:40.675801 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Oct 28 05:17:40.675809 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Oct 28 05:17:40.675816 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Oct 28 05:17:40.675824 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Oct 28 05:17:40.675831 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Oct 28 05:17:40.675839 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Oct 28 05:17:40.675846 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Oct 28 05:17:40.675860 kernel: NX (Execute Disable) protection: active Oct 28 05:17:40.675868 kernel: APIC: Static calls initialized Oct 28 05:17:40.675876 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Oct 28 05:17:40.675884 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Oct 28 05:17:40.675891 kernel: extended physical RAM map: Oct 28 05:17:40.675899 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Oct 28 05:17:40.675906 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Oct 28 05:17:40.675914 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Oct 28 05:17:40.675921 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Oct 28 05:17:40.675929 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Oct 28 05:17:40.675937 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Oct 28 05:17:40.675951 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Oct 28 05:17:40.675958 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Oct 28 05:17:40.675966 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Oct 28 05:17:40.675982 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Oct 28 05:17:40.675996 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Oct 28 05:17:40.676004 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Oct 28 05:17:40.676012 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Oct 28 05:17:40.676020 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Oct 28 05:17:40.676028 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Oct 28 05:17:40.676035 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Oct 28 05:17:40.676043 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Oct 28 05:17:40.676051 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Oct 28 05:17:40.676059 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Oct 28 05:17:40.676073 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Oct 28 05:17:40.676081 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Oct 28 05:17:40.676088 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Oct 28 05:17:40.676096 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Oct 28 05:17:40.676104 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Oct 28 05:17:40.676112 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Oct 28 05:17:40.676120 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Oct 28 05:17:40.676127 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Oct 28 05:17:40.676137 kernel: efi: EFI v2.7 by EDK II Oct 28 05:17:40.676146 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Oct 28 05:17:40.676153 kernel: random: crng init done Oct 28 05:17:40.676237 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Oct 28 05:17:40.676245 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Oct 28 05:17:40.676255 kernel: secureboot: Secure boot disabled Oct 28 05:17:40.676262 kernel: SMBIOS 2.8 present. Oct 28 05:17:40.676270 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Oct 28 05:17:40.676278 kernel: DMI: Memory slots populated: 1/1 Oct 28 05:17:40.676286 kernel: Hypervisor detected: KVM Oct 28 05:17:40.676293 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Oct 28 05:17:40.676301 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Oct 28 05:17:40.676309 kernel: kvm-clock: using sched offset of 5312579861 cycles Oct 28 05:17:40.676318 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Oct 28 05:17:40.676333 kernel: tsc: Detected 2794.748 MHz processor Oct 28 05:17:40.676341 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Oct 28 05:17:40.676349 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Oct 28 05:17:40.676357 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Oct 28 05:17:40.676365 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Oct 28 05:17:40.676374 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Oct 28 05:17:40.676382 kernel: Using GB pages for direct mapping Oct 28 05:17:40.676413 kernel: ACPI: Early table checksum verification disabled Oct 28 05:17:40.676421 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Oct 28 05:17:40.676429 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Oct 28 05:17:40.676437 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Oct 28 05:17:40.676446 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 28 05:17:40.676454 kernel: ACPI: FACS 0x000000009CBDD000 000040 Oct 28 05:17:40.676462 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 28 05:17:40.676478 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 28 05:17:40.676486 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 28 05:17:40.676494 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 28 05:17:40.676502 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Oct 28 05:17:40.676511 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Oct 28 05:17:40.676519 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Oct 28 05:17:40.676535 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Oct 28 05:17:40.676551 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Oct 28 05:17:40.676559 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Oct 28 05:17:40.676567 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Oct 28 05:17:40.676575 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Oct 28 05:17:40.676583 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Oct 28 05:17:40.676591 kernel: No NUMA configuration found Oct 28 05:17:40.676599 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Oct 28 05:17:40.676614 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Oct 28 05:17:40.676622 kernel: Zone ranges: Oct 28 05:17:40.676630 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Oct 28 05:17:40.676638 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Oct 28 05:17:40.676646 kernel: Normal empty Oct 28 05:17:40.676654 kernel: Device empty Oct 28 05:17:40.676662 kernel: Movable zone start for each node Oct 28 05:17:40.676670 kernel: Early memory node ranges Oct 28 05:17:40.676685 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Oct 28 05:17:40.676695 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Oct 28 05:17:40.676703 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Oct 28 05:17:40.676711 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Oct 28 05:17:40.676719 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Oct 28 05:17:40.676727 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Oct 28 05:17:40.676736 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Oct 28 05:17:40.676750 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Oct 28 05:17:40.676760 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Oct 28 05:17:40.676768 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Oct 28 05:17:40.676795 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Oct 28 05:17:40.676809 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Oct 28 05:17:40.676818 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Oct 28 05:17:40.676826 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Oct 28 05:17:40.676834 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Oct 28 05:17:40.676843 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Oct 28 05:17:40.676851 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Oct 28 05:17:40.676866 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Oct 28 05:17:40.676874 kernel: ACPI: PM-Timer IO Port: 0x608 Oct 28 05:17:40.676883 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Oct 28 05:17:40.676891 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Oct 28 05:17:40.676906 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Oct 28 05:17:40.676914 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Oct 28 05:17:40.676923 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Oct 28 05:17:40.676931 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Oct 28 05:17:40.676939 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Oct 28 05:17:40.676948 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Oct 28 05:17:40.676956 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Oct 28 05:17:40.676971 kernel: TSC deadline timer available Oct 28 05:17:40.676980 kernel: CPU topo: Max. logical packages: 1 Oct 28 05:17:40.676988 kernel: CPU topo: Max. logical dies: 1 Oct 28 05:17:40.676996 kernel: CPU topo: Max. dies per package: 1 Oct 28 05:17:40.677005 kernel: CPU topo: Max. threads per core: 1 Oct 28 05:17:40.677013 kernel: CPU topo: Num. cores per package: 4 Oct 28 05:17:40.677021 kernel: CPU topo: Num. threads per package: 4 Oct 28 05:17:40.677030 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Oct 28 05:17:40.677045 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Oct 28 05:17:40.677053 kernel: kvm-guest: KVM setup pv remote TLB flush Oct 28 05:17:40.677061 kernel: kvm-guest: setup PV sched yield Oct 28 05:17:40.677069 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Oct 28 05:17:40.677078 kernel: Booting paravirtualized kernel on KVM Oct 28 05:17:40.677086 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Oct 28 05:17:40.677095 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Oct 28 05:17:40.677110 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Oct 28 05:17:40.677119 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Oct 28 05:17:40.677127 kernel: pcpu-alloc: [0] 0 1 2 3 Oct 28 05:17:40.677135 kernel: kvm-guest: PV spinlocks enabled Oct 28 05:17:40.677143 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Oct 28 05:17:40.677155 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=449db75fd0bf4f00a7b0da93783dc37f82f4a66df937e11c006397de0369495c Oct 28 05:17:40.677175 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Oct 28 05:17:40.677184 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Oct 28 05:17:40.677192 kernel: Fallback order for Node 0: 0 Oct 28 05:17:40.677201 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Oct 28 05:17:40.677209 kernel: Policy zone: DMA32 Oct 28 05:17:40.677217 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Oct 28 05:17:40.677226 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Oct 28 05:17:40.677241 kernel: ftrace: allocating 40092 entries in 157 pages Oct 28 05:17:40.677249 kernel: ftrace: allocated 157 pages with 5 groups Oct 28 05:17:40.677257 kernel: Dynamic Preempt: voluntary Oct 28 05:17:40.677266 kernel: rcu: Preemptible hierarchical RCU implementation. Oct 28 05:17:40.677275 kernel: rcu: RCU event tracing is enabled. Oct 28 05:17:40.677283 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Oct 28 05:17:40.677292 kernel: Trampoline variant of Tasks RCU enabled. Oct 28 05:17:40.677300 kernel: Rude variant of Tasks RCU enabled. Oct 28 05:17:40.677316 kernel: Tracing variant of Tasks RCU enabled. Oct 28 05:17:40.677324 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Oct 28 05:17:40.677333 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Oct 28 05:17:40.677343 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Oct 28 05:17:40.677352 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Oct 28 05:17:40.677361 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Oct 28 05:17:40.677369 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Oct 28 05:17:40.677385 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Oct 28 05:17:40.677404 kernel: Console: colour dummy device 80x25 Oct 28 05:17:40.677413 kernel: printk: legacy console [ttyS0] enabled Oct 28 05:17:40.677421 kernel: ACPI: Core revision 20240827 Oct 28 05:17:40.677429 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Oct 28 05:17:40.677438 kernel: APIC: Switch to symmetric I/O mode setup Oct 28 05:17:40.677446 kernel: x2apic enabled Oct 28 05:17:40.677462 kernel: APIC: Switched APIC routing to: physical x2apic Oct 28 05:17:40.677470 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Oct 28 05:17:40.677479 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Oct 28 05:17:40.677487 kernel: kvm-guest: setup PV IPIs Oct 28 05:17:40.677495 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Oct 28 05:17:40.677504 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Oct 28 05:17:40.677512 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Oct 28 05:17:40.677535 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Oct 28 05:17:40.677545 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Oct 28 05:17:40.677553 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Oct 28 05:17:40.677562 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Oct 28 05:17:40.677570 kernel: Spectre V2 : Mitigation: Retpolines Oct 28 05:17:40.677579 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Oct 28 05:17:40.677587 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Oct 28 05:17:40.677603 kernel: active return thunk: retbleed_return_thunk Oct 28 05:17:40.677612 kernel: RETBleed: Mitigation: untrained return thunk Oct 28 05:17:40.677623 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Oct 28 05:17:40.677631 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Oct 28 05:17:40.677640 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Oct 28 05:17:40.677649 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Oct 28 05:17:40.677657 kernel: active return thunk: srso_return_thunk Oct 28 05:17:40.677673 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Oct 28 05:17:40.677682 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Oct 28 05:17:40.677690 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Oct 28 05:17:40.677698 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Oct 28 05:17:40.677707 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Oct 28 05:17:40.677715 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Oct 28 05:17:40.677723 kernel: Freeing SMP alternatives memory: 32K Oct 28 05:17:40.677739 kernel: pid_max: default: 32768 minimum: 301 Oct 28 05:17:40.677747 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Oct 28 05:17:40.677755 kernel: landlock: Up and running. Oct 28 05:17:40.677763 kernel: SELinux: Initializing. Oct 28 05:17:40.677772 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Oct 28 05:17:40.677780 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Oct 28 05:17:40.677789 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Oct 28 05:17:40.677804 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Oct 28 05:17:40.677813 kernel: ... version: 0 Oct 28 05:17:40.677821 kernel: ... bit width: 48 Oct 28 05:17:40.677829 kernel: ... generic registers: 6 Oct 28 05:17:40.677838 kernel: ... value mask: 0000ffffffffffff Oct 28 05:17:40.677846 kernel: ... max period: 00007fffffffffff Oct 28 05:17:40.677854 kernel: ... fixed-purpose events: 0 Oct 28 05:17:40.677869 kernel: ... event mask: 000000000000003f Oct 28 05:17:40.677878 kernel: signal: max sigframe size: 1776 Oct 28 05:17:40.677886 kernel: rcu: Hierarchical SRCU implementation. Oct 28 05:17:40.677895 kernel: rcu: Max phase no-delay instances is 400. Oct 28 05:17:40.677905 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Oct 28 05:17:40.677914 kernel: smp: Bringing up secondary CPUs ... Oct 28 05:17:40.677922 kernel: smpboot: x86: Booting SMP configuration: Oct 28 05:17:40.677937 kernel: .... node #0, CPUs: #1 #2 #3 Oct 28 05:17:40.677945 kernel: smp: Brought up 1 node, 4 CPUs Oct 28 05:17:40.677954 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Oct 28 05:17:40.677963 kernel: Memory: 2441096K/2565800K available (14336K kernel code, 2443K rwdata, 29892K rodata, 15960K init, 2084K bss, 118764K reserved, 0K cma-reserved) Oct 28 05:17:40.677972 kernel: devtmpfs: initialized Oct 28 05:17:40.677980 kernel: x86/mm: Memory block size: 128MB Oct 28 05:17:40.677989 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Oct 28 05:17:40.678004 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Oct 28 05:17:40.678013 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Oct 28 05:17:40.678022 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Oct 28 05:17:40.678030 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Oct 28 05:17:40.678039 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Oct 28 05:17:40.678047 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Oct 28 05:17:40.678056 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Oct 28 05:17:40.678071 kernel: pinctrl core: initialized pinctrl subsystem Oct 28 05:17:40.678079 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Oct 28 05:17:40.678087 kernel: audit: initializing netlink subsys (disabled) Oct 28 05:17:40.678096 kernel: audit: type=2000 audit(1761628656.777:1): state=initialized audit_enabled=0 res=1 Oct 28 05:17:40.678104 kernel: thermal_sys: Registered thermal governor 'step_wise' Oct 28 05:17:40.678113 kernel: thermal_sys: Registered thermal governor 'user_space' Oct 28 05:17:40.678121 kernel: cpuidle: using governor menu Oct 28 05:17:40.678136 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Oct 28 05:17:40.678144 kernel: dca service started, version 1.12.1 Oct 28 05:17:40.678152 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Oct 28 05:17:40.678161 kernel: PCI: Using configuration type 1 for base access Oct 28 05:17:40.678169 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Oct 28 05:17:40.678178 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Oct 28 05:17:40.678186 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Oct 28 05:17:40.678201 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Oct 28 05:17:40.678210 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Oct 28 05:17:40.678218 kernel: ACPI: Added _OSI(Module Device) Oct 28 05:17:40.678227 kernel: ACPI: Added _OSI(Processor Device) Oct 28 05:17:40.678235 kernel: ACPI: Added _OSI(Processor Aggregator Device) Oct 28 05:17:40.678243 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Oct 28 05:17:40.678252 kernel: ACPI: Interpreter enabled Oct 28 05:17:40.678266 kernel: ACPI: PM: (supports S0 S3 S5) Oct 28 05:17:40.678275 kernel: ACPI: Using IOAPIC for interrupt routing Oct 28 05:17:40.678283 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Oct 28 05:17:40.678292 kernel: PCI: Using E820 reservations for host bridge windows Oct 28 05:17:40.678300 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Oct 28 05:17:40.678308 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Oct 28 05:17:40.678573 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Oct 28 05:17:40.678793 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Oct 28 05:17:40.678989 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Oct 28 05:17:40.679006 kernel: PCI host bridge to bus 0000:00 Oct 28 05:17:40.679197 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Oct 28 05:17:40.679361 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Oct 28 05:17:40.679575 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Oct 28 05:17:40.679736 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Oct 28 05:17:40.679895 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Oct 28 05:17:40.680055 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Oct 28 05:17:40.680215 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Oct 28 05:17:40.680429 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Oct 28 05:17:40.680672 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Oct 28 05:17:40.680850 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Oct 28 05:17:40.681039 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Oct 28 05:17:40.681212 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Oct 28 05:17:40.681383 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Oct 28 05:17:40.681610 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Oct 28 05:17:40.681802 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Oct 28 05:17:40.681975 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Oct 28 05:17:40.682147 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Oct 28 05:17:40.682331 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Oct 28 05:17:40.682534 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Oct 28 05:17:40.682725 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Oct 28 05:17:40.682899 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Oct 28 05:17:40.683083 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Oct 28 05:17:40.683257 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Oct 28 05:17:40.683459 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Oct 28 05:17:40.683653 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Oct 28 05:17:40.683846 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Oct 28 05:17:40.684026 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Oct 28 05:17:40.684199 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Oct 28 05:17:40.684381 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Oct 28 05:17:40.684598 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Oct 28 05:17:40.684787 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Oct 28 05:17:40.684970 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Oct 28 05:17:40.685142 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Oct 28 05:17:40.685154 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Oct 28 05:17:40.685163 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Oct 28 05:17:40.685171 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Oct 28 05:17:40.685180 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Oct 28 05:17:40.685200 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Oct 28 05:17:40.685209 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Oct 28 05:17:40.685217 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Oct 28 05:17:40.685225 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Oct 28 05:17:40.685234 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Oct 28 05:17:40.685242 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Oct 28 05:17:40.685250 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Oct 28 05:17:40.685267 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Oct 28 05:17:40.685275 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Oct 28 05:17:40.685283 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Oct 28 05:17:40.685292 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Oct 28 05:17:40.685300 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Oct 28 05:17:40.685309 kernel: iommu: Default domain type: Translated Oct 28 05:17:40.685317 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Oct 28 05:17:40.685332 kernel: efivars: Registered efivars operations Oct 28 05:17:40.685341 kernel: PCI: Using ACPI for IRQ routing Oct 28 05:17:40.685349 kernel: PCI: pci_cache_line_size set to 64 bytes Oct 28 05:17:40.685357 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Oct 28 05:17:40.685366 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Oct 28 05:17:40.685374 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Oct 28 05:17:40.685382 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Oct 28 05:17:40.685411 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Oct 28 05:17:40.685420 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Oct 28 05:17:40.685428 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Oct 28 05:17:40.685437 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Oct 28 05:17:40.685626 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Oct 28 05:17:40.685799 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Oct 28 05:17:40.685987 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Oct 28 05:17:40.685998 kernel: vgaarb: loaded Oct 28 05:17:40.686007 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Oct 28 05:17:40.686016 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Oct 28 05:17:40.686025 kernel: clocksource: Switched to clocksource kvm-clock Oct 28 05:17:40.686033 kernel: VFS: Disk quotas dquot_6.6.0 Oct 28 05:17:40.686042 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Oct 28 05:17:40.686050 kernel: pnp: PnP ACPI init Oct 28 05:17:40.686291 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Oct 28 05:17:40.686320 kernel: pnp: PnP ACPI: found 6 devices Oct 28 05:17:40.686332 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Oct 28 05:17:40.686344 kernel: NET: Registered PF_INET protocol family Oct 28 05:17:40.686361 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Oct 28 05:17:40.686370 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Oct 28 05:17:40.686386 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Oct 28 05:17:40.686409 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Oct 28 05:17:40.686418 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Oct 28 05:17:40.686427 kernel: TCP: Hash tables configured (established 32768 bind 32768) Oct 28 05:17:40.686436 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Oct 28 05:17:40.686445 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Oct 28 05:17:40.686453 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Oct 28 05:17:40.686470 kernel: NET: Registered PF_XDP protocol family Oct 28 05:17:40.686663 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Oct 28 05:17:40.686839 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Oct 28 05:17:40.687002 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Oct 28 05:17:40.687162 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Oct 28 05:17:40.687325 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Oct 28 05:17:40.687534 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Oct 28 05:17:40.687700 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Oct 28 05:17:40.687859 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Oct 28 05:17:40.687871 kernel: PCI: CLS 0 bytes, default 64 Oct 28 05:17:40.687880 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Oct 28 05:17:40.687902 kernel: Initialise system trusted keyrings Oct 28 05:17:40.687911 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Oct 28 05:17:40.687920 kernel: Key type asymmetric registered Oct 28 05:17:40.687929 kernel: Asymmetric key parser 'x509' registered Oct 28 05:17:40.687937 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Oct 28 05:17:40.687946 kernel: io scheduler mq-deadline registered Oct 28 05:17:40.687963 kernel: io scheduler kyber registered Oct 28 05:17:40.687971 kernel: io scheduler bfq registered Oct 28 05:17:40.687980 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Oct 28 05:17:40.687989 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Oct 28 05:17:40.687998 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Oct 28 05:17:40.688007 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Oct 28 05:17:40.688016 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Oct 28 05:17:40.688025 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Oct 28 05:17:40.688041 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Oct 28 05:17:40.688050 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Oct 28 05:17:40.688059 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Oct 28 05:17:40.688238 kernel: rtc_cmos 00:04: RTC can wake from S4 Oct 28 05:17:40.688252 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Oct 28 05:17:40.688452 kernel: rtc_cmos 00:04: registered as rtc0 Oct 28 05:17:40.688650 kernel: rtc_cmos 00:04: setting system clock to 2025-10-28T05:17:38 UTC (1761628658) Oct 28 05:17:40.688818 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Oct 28 05:17:40.688829 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Oct 28 05:17:40.688838 kernel: efifb: probing for efifb Oct 28 05:17:40.688847 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Oct 28 05:17:40.688856 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Oct 28 05:17:40.688864 kernel: efifb: scrolling: redraw Oct 28 05:17:40.688884 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Oct 28 05:17:40.688893 kernel: Console: switching to colour frame buffer device 160x50 Oct 28 05:17:40.688902 kernel: fb0: EFI VGA frame buffer device Oct 28 05:17:40.688910 kernel: pstore: Using crash dump compression: deflate Oct 28 05:17:40.688919 kernel: pstore: Registered efi_pstore as persistent store backend Oct 28 05:17:40.688928 kernel: NET: Registered PF_INET6 protocol family Oct 28 05:17:40.688936 kernel: Segment Routing with IPv6 Oct 28 05:17:40.688952 kernel: In-situ OAM (IOAM) with IPv6 Oct 28 05:17:40.688961 kernel: NET: Registered PF_PACKET protocol family Oct 28 05:17:40.688970 kernel: Key type dns_resolver registered Oct 28 05:17:40.688979 kernel: IPI shorthand broadcast: enabled Oct 28 05:17:40.688987 kernel: sched_clock: Marking stable (1490040346, 306465926)->(1943492672, -146986400) Oct 28 05:17:40.689002 kernel: registered taskstats version 1 Oct 28 05:17:40.689011 kernel: Loading compiled-in X.509 certificates Oct 28 05:17:40.689026 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.54-flatcar: a9d98af1927e389c63ed03bf44a9f2758bf88a8e' Oct 28 05:17:40.689035 kernel: Demotion targets for Node 0: null Oct 28 05:17:40.689044 kernel: Key type .fscrypt registered Oct 28 05:17:40.689052 kernel: Key type fscrypt-provisioning registered Oct 28 05:17:40.689061 kernel: ima: No TPM chip found, activating TPM-bypass! Oct 28 05:17:40.689070 kernel: ima: Allocated hash algorithm: sha1 Oct 28 05:17:40.689079 kernel: ima: No architecture policies found Oct 28 05:17:40.689087 kernel: clk: Disabling unused clocks Oct 28 05:17:40.689103 kernel: Freeing unused kernel image (initmem) memory: 15960K Oct 28 05:17:40.689112 kernel: Write protecting the kernel read-only data: 45056k Oct 28 05:17:40.689121 kernel: Freeing unused kernel image (rodata/data gap) memory: 828K Oct 28 05:17:40.689130 kernel: Run /init as init process Oct 28 05:17:40.689138 kernel: with arguments: Oct 28 05:17:40.689147 kernel: /init Oct 28 05:17:40.689156 kernel: with environment: Oct 28 05:17:40.689171 kernel: HOME=/ Oct 28 05:17:40.689179 kernel: TERM=linux Oct 28 05:17:40.689188 kernel: SCSI subsystem initialized Oct 28 05:17:40.689197 kernel: libata version 3.00 loaded. Oct 28 05:17:40.689376 kernel: ahci 0000:00:1f.2: version 3.0 Oct 28 05:17:40.689388 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Oct 28 05:17:40.689652 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Oct 28 05:17:40.689845 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Oct 28 05:17:40.690020 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Oct 28 05:17:40.690219 kernel: scsi host0: ahci Oct 28 05:17:40.690438 kernel: scsi host1: ahci Oct 28 05:17:40.690639 kernel: scsi host2: ahci Oct 28 05:17:40.690843 kernel: scsi host3: ahci Oct 28 05:17:40.691028 kernel: scsi host4: ahci Oct 28 05:17:40.691212 kernel: scsi host5: ahci Oct 28 05:17:40.691225 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 26 lpm-pol 1 Oct 28 05:17:40.691234 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 26 lpm-pol 1 Oct 28 05:17:40.691243 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 26 lpm-pol 1 Oct 28 05:17:40.691264 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 26 lpm-pol 1 Oct 28 05:17:40.691273 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 26 lpm-pol 1 Oct 28 05:17:40.691282 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 26 lpm-pol 1 Oct 28 05:17:40.691291 kernel: ata6: SATA link down (SStatus 0 SControl 300) Oct 28 05:17:40.691299 kernel: ata5: SATA link down (SStatus 0 SControl 300) Oct 28 05:17:40.691308 kernel: ata1: SATA link down (SStatus 0 SControl 300) Oct 28 05:17:40.691317 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Oct 28 05:17:40.691334 kernel: ata2: SATA link down (SStatus 0 SControl 300) Oct 28 05:17:40.691342 kernel: ata4: SATA link down (SStatus 0 SControl 300) Oct 28 05:17:40.691351 kernel: ata3.00: LPM support broken, forcing max_power Oct 28 05:17:40.691359 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Oct 28 05:17:40.691368 kernel: ata3.00: applying bridge limits Oct 28 05:17:40.691377 kernel: ata3.00: LPM support broken, forcing max_power Oct 28 05:17:40.691385 kernel: ata3.00: configured for UDMA/100 Oct 28 05:17:40.695353 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Oct 28 05:17:40.695592 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Oct 28 05:17:40.695780 kernel: virtio_blk virtio1: [vda] 27000832 512-byte logical blocks (13.8 GB/12.9 GiB) Oct 28 05:17:40.695794 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Oct 28 05:17:40.695803 kernel: GPT:16515071 != 27000831 Oct 28 05:17:40.695812 kernel: GPT:Alternate GPT header not at the end of the disk. Oct 28 05:17:40.695841 kernel: GPT:16515071 != 27000831 Oct 28 05:17:40.695849 kernel: GPT: Use GNU Parted to correct GPT errors. Oct 28 05:17:40.695858 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 28 05:17:40.695867 kernel: Invalid ELF header magic: != \u007fELF Oct 28 05:17:40.696073 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Oct 28 05:17:40.696086 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Oct 28 05:17:40.696276 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Oct 28 05:17:40.696299 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Oct 28 05:17:40.696308 kernel: device-mapper: uevent: version 1.0.3 Oct 28 05:17:40.696317 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Oct 28 05:17:40.696326 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Oct 28 05:17:40.696335 kernel: Invalid ELF header magic: != \u007fELF Oct 28 05:17:40.696344 kernel: Invalid ELF header magic: != \u007fELF Oct 28 05:17:40.696352 kernel: raid6: avx2x4 gen() 28308 MB/s Oct 28 05:17:40.696369 kernel: raid6: avx2x2 gen() 27628 MB/s Oct 28 05:17:40.696378 kernel: raid6: avx2x1 gen() 24531 MB/s Oct 28 05:17:40.696386 kernel: raid6: using algorithm avx2x4 gen() 28308 MB/s Oct 28 05:17:40.696409 kernel: raid6: .... xor() 8173 MB/s, rmw enabled Oct 28 05:17:40.696418 kernel: raid6: using avx2x2 recovery algorithm Oct 28 05:17:40.696437 kernel: Invalid ELF header magic: != \u007fELF Oct 28 05:17:40.696446 kernel: Invalid ELF header magic: != \u007fELF Oct 28 05:17:40.696454 kernel: Invalid ELF header magic: != \u007fELF Oct 28 05:17:40.696478 kernel: xor: automatically using best checksumming function avx Oct 28 05:17:40.698273 kernel: Invalid ELF header magic: != \u007fELF Oct 28 05:17:40.698284 kernel: Btrfs loaded, zoned=no, fsverity=no Oct 28 05:17:40.698294 kernel: BTRFS: device fsid 98ad3ab2-0171-42ae-a5fc-7be2369f5a89 devid 1 transid 37 /dev/mapper/usr (253:0) scanned by mount (175) Oct 28 05:17:40.698303 kernel: BTRFS info (device dm-0): first mount of filesystem 98ad3ab2-0171-42ae-a5fc-7be2369f5a89 Oct 28 05:17:40.698312 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Oct 28 05:17:40.698321 kernel: BTRFS info (device dm-0): disabling log replay at mount time Oct 28 05:17:40.698344 kernel: BTRFS info (device dm-0): enabling free space tree Oct 28 05:17:40.698353 kernel: Invalid ELF header magic: != \u007fELF Oct 28 05:17:40.698361 kernel: loop: module loaded Oct 28 05:17:40.698370 kernel: loop0: detected capacity change from 0 to 100136 Oct 28 05:17:40.698379 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Oct 28 05:17:40.698389 systemd[1]: Successfully made /usr/ read-only. Oct 28 05:17:40.698414 systemd[1]: systemd 257.7 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Oct 28 05:17:40.698432 systemd[1]: Detected virtualization kvm. Oct 28 05:17:40.698441 systemd[1]: Detected architecture x86-64. Oct 28 05:17:40.698451 systemd[1]: Running in initrd. Oct 28 05:17:40.698460 systemd[1]: No hostname configured, using default hostname. Oct 28 05:17:40.698470 systemd[1]: Hostname set to . Oct 28 05:17:40.698479 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Oct 28 05:17:40.698495 systemd[1]: Queued start job for default target initrd.target. Oct 28 05:17:40.698504 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Oct 28 05:17:40.698514 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 28 05:17:40.698533 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 28 05:17:40.698546 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Oct 28 05:17:40.698557 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 28 05:17:40.698575 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Oct 28 05:17:40.698585 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Oct 28 05:17:40.698595 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 28 05:17:40.698604 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 28 05:17:40.698613 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Oct 28 05:17:40.698623 systemd[1]: Reached target paths.target - Path Units. Oct 28 05:17:40.698639 systemd[1]: Reached target slices.target - Slice Units. Oct 28 05:17:40.698648 systemd[1]: Reached target swap.target - Swaps. Oct 28 05:17:40.698658 systemd[1]: Reached target timers.target - Timer Units. Oct 28 05:17:40.698667 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Oct 28 05:17:40.698676 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 28 05:17:40.698686 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 28 05:17:40.698695 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Oct 28 05:17:40.698711 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 28 05:17:40.698721 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 28 05:17:40.698731 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 28 05:17:40.698740 systemd[1]: Reached target sockets.target - Socket Units. Oct 28 05:17:40.698749 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Oct 28 05:17:40.698759 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Oct 28 05:17:40.699653 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 28 05:17:40.699665 systemd[1]: Finished network-cleanup.service - Network Cleanup. Oct 28 05:17:40.699675 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Oct 28 05:17:40.699684 systemd[1]: Starting systemd-fsck-usr.service... Oct 28 05:17:40.699702 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 28 05:17:40.699712 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 28 05:17:40.699722 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 28 05:17:40.699739 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Oct 28 05:17:40.699748 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 28 05:17:40.699757 systemd[1]: Finished systemd-fsck-usr.service. Oct 28 05:17:40.699767 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 28 05:17:40.699817 systemd-journald[310]: Collecting audit messages is disabled. Oct 28 05:17:40.699839 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 28 05:17:40.699849 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 28 05:17:40.699866 systemd-journald[310]: Journal started Oct 28 05:17:40.699885 systemd-journald[310]: Runtime Journal (/run/log/journal/10408c3878b84479b95d9d00a690fd2b) is 6M, max 48.1M, 42.1M free. Oct 28 05:17:40.702988 systemd[1]: Started systemd-journald.service - Journal Service. Oct 28 05:17:40.709429 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Oct 28 05:17:40.713187 systemd-modules-load[312]: Inserted module 'br_netfilter' Oct 28 05:17:40.715822 kernel: Bridge firewalling registered Oct 28 05:17:40.713500 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 28 05:17:40.719060 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 28 05:17:40.724124 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 28 05:17:40.728644 systemd-tmpfiles[330]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Oct 28 05:17:40.732508 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 28 05:17:40.737335 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 28 05:17:40.742454 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 28 05:17:40.757727 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 28 05:17:40.771492 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 28 05:17:40.773658 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 28 05:17:40.780183 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 28 05:17:40.786822 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Oct 28 05:17:40.809889 dracut-cmdline[355]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=449db75fd0bf4f00a7b0da93783dc37f82f4a66df937e11c006397de0369495c Oct 28 05:17:40.846029 systemd-resolved[350]: Positive Trust Anchors: Oct 28 05:17:40.846048 systemd-resolved[350]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 28 05:17:40.846054 systemd-resolved[350]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Oct 28 05:17:40.846095 systemd-resolved[350]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 28 05:17:40.874062 systemd-resolved[350]: Defaulting to hostname 'linux'. Oct 28 05:17:40.876653 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 28 05:17:40.878644 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 28 05:17:40.950468 kernel: Loading iSCSI transport class v2.0-870. Oct 28 05:17:40.968475 kernel: iscsi: registered transport (tcp) Oct 28 05:17:41.024078 kernel: iscsi: registered transport (qla4xxx) Oct 28 05:17:41.024265 kernel: QLogic iSCSI HBA Driver Oct 28 05:17:41.053341 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 28 05:17:41.090969 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 28 05:17:41.092373 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 28 05:17:41.162632 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Oct 28 05:17:41.164464 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Oct 28 05:17:41.169356 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Oct 28 05:17:41.217377 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Oct 28 05:17:41.219332 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 28 05:17:41.251292 systemd-udevd[591]: Using default interface naming scheme 'v257'. Oct 28 05:17:41.267999 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 28 05:17:41.273619 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Oct 28 05:17:41.308924 dracut-pre-trigger[651]: rd.md=0: removing MD RAID activation Oct 28 05:17:41.320939 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 28 05:17:41.327330 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 28 05:17:41.344472 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Oct 28 05:17:41.345995 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 28 05:17:41.389797 systemd-networkd[714]: lo: Link UP Oct 28 05:17:41.389805 systemd-networkd[714]: lo: Gained carrier Oct 28 05:17:41.390433 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 28 05:17:41.393034 systemd[1]: Reached target network.target - Network. Oct 28 05:17:41.451526 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 28 05:17:41.455009 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Oct 28 05:17:41.592885 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Oct 28 05:17:41.613087 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Oct 28 05:17:41.628426 kernel: cryptd: max_cpu_qlen set to 1000 Oct 28 05:17:41.636374 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 28 05:17:41.646995 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Oct 28 05:17:41.655425 kernel: AES CTR mode by8 optimization enabled Oct 28 05:17:41.656572 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Oct 28 05:17:41.667461 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Oct 28 05:17:41.680019 systemd-networkd[714]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Oct 28 05:17:41.680031 systemd-networkd[714]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 28 05:17:41.687685 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 28 05:17:41.687887 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 28 05:17:41.697852 disk-uuid[815]: Primary Header is updated. Oct 28 05:17:41.697852 disk-uuid[815]: Secondary Entries is updated. Oct 28 05:17:41.697852 disk-uuid[815]: Secondary Header is updated. Oct 28 05:17:41.691794 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Oct 28 05:17:41.692694 systemd-networkd[714]: eth0: Link UP Oct 28 05:17:41.692909 systemd-networkd[714]: eth0: Gained carrier Oct 28 05:17:41.692921 systemd-networkd[714]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Oct 28 05:17:41.698341 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 28 05:17:41.710458 systemd-networkd[714]: eth0: DHCPv4 address 10.0.0.103/16, gateway 10.0.0.1 acquired from 10.0.0.1 Oct 28 05:17:41.725452 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 28 05:17:41.725679 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 28 05:17:41.756536 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 28 05:17:41.816153 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Oct 28 05:17:41.818845 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Oct 28 05:17:41.819739 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 28 05:17:41.820267 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 28 05:17:41.821921 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Oct 28 05:17:41.841757 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 28 05:17:41.858014 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Oct 28 05:17:41.993912 systemd-resolved[350]: Detected conflict on linux IN A 10.0.0.103 Oct 28 05:17:41.993945 systemd-resolved[350]: Hostname conflict, changing published hostname from 'linux' to 'linux11'. Oct 28 05:17:42.751131 disk-uuid[819]: Warning: The kernel is still using the old partition table. Oct 28 05:17:42.751131 disk-uuid[819]: The new table will be used at the next reboot or after you Oct 28 05:17:42.751131 disk-uuid[819]: run partprobe(8) or kpartx(8) Oct 28 05:17:42.751131 disk-uuid[819]: The operation has completed successfully. Oct 28 05:17:42.766909 systemd[1]: disk-uuid.service: Deactivated successfully. Oct 28 05:17:42.767109 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Oct 28 05:17:42.772554 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Oct 28 05:17:42.806426 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (860) Oct 28 05:17:42.806461 kernel: BTRFS info (device vda6): first mount of filesystem 7acd037c-32ce-4796-90d6-101869832417 Oct 28 05:17:42.809376 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 28 05:17:42.812432 kernel: BTRFS info (device vda6): turning on async discard Oct 28 05:17:42.812456 kernel: BTRFS info (device vda6): enabling free space tree Oct 28 05:17:42.821417 kernel: BTRFS info (device vda6): last unmount of filesystem 7acd037c-32ce-4796-90d6-101869832417 Oct 28 05:17:42.822415 systemd[1]: Finished ignition-setup.service - Ignition (setup). Oct 28 05:17:42.823778 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Oct 28 05:17:42.943798 ignition[879]: Ignition 2.22.0 Oct 28 05:17:42.943813 ignition[879]: Stage: fetch-offline Oct 28 05:17:42.943863 ignition[879]: no configs at "/usr/lib/ignition/base.d" Oct 28 05:17:42.943881 ignition[879]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 28 05:17:42.943983 ignition[879]: parsed url from cmdline: "" Oct 28 05:17:42.943989 ignition[879]: no config URL provided Oct 28 05:17:42.943996 ignition[879]: reading system config file "/usr/lib/ignition/user.ign" Oct 28 05:17:42.944012 ignition[879]: no config at "/usr/lib/ignition/user.ign" Oct 28 05:17:42.944065 ignition[879]: op(1): [started] loading QEMU firmware config module Oct 28 05:17:42.944072 ignition[879]: op(1): executing: "modprobe" "qemu_fw_cfg" Oct 28 05:17:42.954451 ignition[879]: op(1): [finished] loading QEMU firmware config module Oct 28 05:17:43.035449 ignition[879]: parsing config with SHA512: e88b04dcb56cb4b0a50a592f0c0c420b12e891dd5d4456543f95fc682d646ee2ec39677cb9611b858ee780149084bef4e6d84b1a633c47c8ef64e0df5ddb1bd7 Oct 28 05:17:43.041120 unknown[879]: fetched base config from "system" Oct 28 05:17:43.041449 unknown[879]: fetched user config from "qemu" Oct 28 05:17:43.041959 ignition[879]: fetch-offline: fetch-offline passed Oct 28 05:17:43.042056 ignition[879]: Ignition finished successfully Oct 28 05:17:43.049509 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Oct 28 05:17:43.049794 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Oct 28 05:17:43.050733 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Oct 28 05:17:43.087696 ignition[889]: Ignition 2.22.0 Oct 28 05:17:43.087710 ignition[889]: Stage: kargs Oct 28 05:17:43.087852 ignition[889]: no configs at "/usr/lib/ignition/base.d" Oct 28 05:17:43.087863 ignition[889]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 28 05:17:43.088878 ignition[889]: kargs: kargs passed Oct 28 05:17:43.088949 ignition[889]: Ignition finished successfully Oct 28 05:17:43.095561 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Oct 28 05:17:43.098798 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Oct 28 05:17:43.139184 ignition[897]: Ignition 2.22.0 Oct 28 05:17:43.139200 ignition[897]: Stage: disks Oct 28 05:17:43.139345 ignition[897]: no configs at "/usr/lib/ignition/base.d" Oct 28 05:17:43.139355 ignition[897]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 28 05:17:43.140379 ignition[897]: disks: disks passed Oct 28 05:17:43.140483 ignition[897]: Ignition finished successfully Oct 28 05:17:43.146501 systemd[1]: Finished ignition-disks.service - Ignition (disks). Oct 28 05:17:43.149449 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Oct 28 05:17:43.152651 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 28 05:17:43.153005 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 28 05:17:43.158157 systemd[1]: Reached target sysinit.target - System Initialization. Oct 28 05:17:43.158973 systemd[1]: Reached target basic.target - Basic System. Oct 28 05:17:43.165361 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Oct 28 05:17:43.212127 systemd-fsck[907]: ROOT: clean, 15/456736 files, 38230/456704 blocks Oct 28 05:17:43.221951 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Oct 28 05:17:43.227353 systemd[1]: Mounting sysroot.mount - /sysroot... Oct 28 05:17:43.545463 kernel: EXT4-fs (vda9): mounted filesystem 0ce42fa0-8451-4928-b788-6e54ab295d7a r/w with ordered data mode. Quota mode: none. Oct 28 05:17:43.546979 systemd[1]: Mounted sysroot.mount - /sysroot. Oct 28 05:17:43.550269 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Oct 28 05:17:43.555288 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 28 05:17:43.558963 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Oct 28 05:17:43.562065 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Oct 28 05:17:43.562101 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Oct 28 05:17:43.562128 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Oct 28 05:17:43.592750 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Oct 28 05:17:43.596950 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Oct 28 05:17:43.601278 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (916) Oct 28 05:17:43.603423 kernel: BTRFS info (device vda6): first mount of filesystem 7acd037c-32ce-4796-90d6-101869832417 Oct 28 05:17:43.603457 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 28 05:17:43.608796 kernel: BTRFS info (device vda6): turning on async discard Oct 28 05:17:43.608825 kernel: BTRFS info (device vda6): enabling free space tree Oct 28 05:17:43.610228 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 28 05:17:43.696158 initrd-setup-root[940]: cut: /sysroot/etc/passwd: No such file or directory Oct 28 05:17:43.702706 initrd-setup-root[947]: cut: /sysroot/etc/group: No such file or directory Oct 28 05:17:43.712196 initrd-setup-root[954]: cut: /sysroot/etc/shadow: No such file or directory Oct 28 05:17:43.717837 initrd-setup-root[961]: cut: /sysroot/etc/gshadow: No such file or directory Oct 28 05:17:43.760928 systemd-networkd[714]: eth0: Gained IPv6LL Oct 28 05:17:43.818368 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Oct 28 05:17:43.822978 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Oct 28 05:17:43.825681 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Oct 28 05:17:43.893236 systemd[1]: sysroot-oem.mount: Deactivated successfully. Oct 28 05:17:43.917599 kernel: BTRFS info (device vda6): last unmount of filesystem 7acd037c-32ce-4796-90d6-101869832417 Oct 28 05:17:43.934559 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Oct 28 05:17:43.956201 ignition[1029]: INFO : Ignition 2.22.0 Oct 28 05:17:43.956201 ignition[1029]: INFO : Stage: mount Oct 28 05:17:43.958727 ignition[1029]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 28 05:17:43.958727 ignition[1029]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 28 05:17:43.958727 ignition[1029]: INFO : mount: mount passed Oct 28 05:17:43.958727 ignition[1029]: INFO : Ignition finished successfully Oct 28 05:17:43.967162 systemd[1]: Finished ignition-mount.service - Ignition (mount). Oct 28 05:17:43.968371 systemd[1]: Starting ignition-files.service - Ignition (files)... Oct 28 05:17:43.996924 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 28 05:17:44.011485 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1042) Oct 28 05:17:44.011537 kernel: BTRFS info (device vda6): first mount of filesystem 7acd037c-32ce-4796-90d6-101869832417 Oct 28 05:17:44.011553 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 28 05:17:44.016739 kernel: BTRFS info (device vda6): turning on async discard Oct 28 05:17:44.016762 kernel: BTRFS info (device vda6): enabling free space tree Oct 28 05:17:44.018667 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 28 05:17:44.073882 ignition[1059]: INFO : Ignition 2.22.0 Oct 28 05:17:44.073882 ignition[1059]: INFO : Stage: files Oct 28 05:17:44.076585 ignition[1059]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 28 05:17:44.076585 ignition[1059]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 28 05:17:44.080470 ignition[1059]: DEBUG : files: compiled without relabeling support, skipping Oct 28 05:17:44.082706 ignition[1059]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Oct 28 05:17:44.082706 ignition[1059]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Oct 28 05:17:44.091241 ignition[1059]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Oct 28 05:17:44.093733 ignition[1059]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Oct 28 05:17:44.096191 ignition[1059]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Oct 28 05:17:44.096191 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Oct 28 05:17:44.096191 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Oct 28 05:17:44.094281 unknown[1059]: wrote ssh authorized keys file for user: core Oct 28 05:17:44.143896 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Oct 28 05:17:44.339346 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Oct 28 05:17:44.339346 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Oct 28 05:17:44.345837 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Oct 28 05:17:44.345837 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Oct 28 05:17:44.345837 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Oct 28 05:17:44.345837 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 28 05:17:44.345837 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 28 05:17:44.345837 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 28 05:17:44.345837 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 28 05:17:44.580102 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Oct 28 05:17:44.590458 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Oct 28 05:17:44.590458 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 28 05:17:44.793379 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 28 05:17:44.793379 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 28 05:17:44.806692 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Oct 28 05:17:45.240261 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Oct 28 05:17:45.890774 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 28 05:17:45.890774 ignition[1059]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Oct 28 05:17:45.896977 ignition[1059]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 28 05:17:45.984344 ignition[1059]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 28 05:17:45.984344 ignition[1059]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Oct 28 05:17:45.984344 ignition[1059]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Oct 28 05:17:45.993043 ignition[1059]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Oct 28 05:17:45.993043 ignition[1059]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Oct 28 05:17:45.993043 ignition[1059]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Oct 28 05:17:45.993043 ignition[1059]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Oct 28 05:17:46.014232 ignition[1059]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Oct 28 05:17:46.049510 ignition[1059]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Oct 28 05:17:46.052859 ignition[1059]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Oct 28 05:17:46.052859 ignition[1059]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Oct 28 05:17:46.052859 ignition[1059]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Oct 28 05:17:46.052859 ignition[1059]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Oct 28 05:17:46.052859 ignition[1059]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Oct 28 05:17:46.052859 ignition[1059]: INFO : files: files passed Oct 28 05:17:46.052859 ignition[1059]: INFO : Ignition finished successfully Oct 28 05:17:46.063717 systemd[1]: Finished ignition-files.service - Ignition (files). Oct 28 05:17:46.067639 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Oct 28 05:17:46.076095 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Oct 28 05:17:46.096020 systemd[1]: ignition-quench.service: Deactivated successfully. Oct 28 05:17:46.096152 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Oct 28 05:17:46.104489 initrd-setup-root-after-ignition[1090]: grep: /sysroot/oem/oem-release: No such file or directory Oct 28 05:17:46.109300 initrd-setup-root-after-ignition[1092]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 28 05:17:46.112172 initrd-setup-root-after-ignition[1092]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Oct 28 05:17:46.126582 initrd-setup-root-after-ignition[1096]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 28 05:17:46.127996 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 28 05:17:46.131573 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Oct 28 05:17:46.137113 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Oct 28 05:17:46.232194 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Oct 28 05:17:46.232376 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Oct 28 05:17:46.234615 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Oct 28 05:17:46.239614 systemd[1]: Reached target initrd.target - Initrd Default Target. Oct 28 05:17:46.240568 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Oct 28 05:17:46.242038 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Oct 28 05:17:46.282956 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 28 05:17:46.286545 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Oct 28 05:17:46.317168 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Oct 28 05:17:46.317566 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Oct 28 05:17:46.319612 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 28 05:17:46.324962 systemd[1]: Stopped target timers.target - Timer Units. Oct 28 05:17:46.329501 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Oct 28 05:17:46.329676 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 28 05:17:46.335042 systemd[1]: Stopped target initrd.target - Initrd Default Target. Oct 28 05:17:46.336830 systemd[1]: Stopped target basic.target - Basic System. Oct 28 05:17:46.338530 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Oct 28 05:17:46.339303 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Oct 28 05:17:46.346027 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Oct 28 05:17:46.349661 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Oct 28 05:17:46.351464 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Oct 28 05:17:46.356322 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Oct 28 05:17:46.358033 systemd[1]: Stopped target sysinit.target - System Initialization. Oct 28 05:17:46.358846 systemd[1]: Stopped target local-fs.target - Local File Systems. Oct 28 05:17:46.359380 systemd[1]: Stopped target swap.target - Swaps. Oct 28 05:17:46.368639 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Oct 28 05:17:46.368813 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Oct 28 05:17:46.374774 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Oct 28 05:17:46.374981 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 28 05:17:46.375752 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Oct 28 05:17:46.375965 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 28 05:17:46.381753 systemd[1]: dracut-initqueue.service: Deactivated successfully. Oct 28 05:17:46.381914 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Oct 28 05:17:46.389009 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Oct 28 05:17:46.389167 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Oct 28 05:17:46.390742 systemd[1]: Stopped target paths.target - Path Units. Oct 28 05:17:46.391187 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Oct 28 05:17:46.401808 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 28 05:17:46.402031 systemd[1]: Stopped target slices.target - Slice Units. Oct 28 05:17:46.407827 systemd[1]: Stopped target sockets.target - Socket Units. Oct 28 05:17:46.409424 systemd[1]: iscsid.socket: Deactivated successfully. Oct 28 05:17:46.409580 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Oct 28 05:17:46.412354 systemd[1]: iscsiuio.socket: Deactivated successfully. Oct 28 05:17:46.412557 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 28 05:17:46.416982 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Oct 28 05:17:46.417218 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 28 05:17:46.420571 systemd[1]: ignition-files.service: Deactivated successfully. Oct 28 05:17:46.420746 systemd[1]: Stopped ignition-files.service - Ignition (files). Oct 28 05:17:46.426874 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Oct 28 05:17:46.431013 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Oct 28 05:17:46.433087 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Oct 28 05:17:46.433295 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 28 05:17:46.436783 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Oct 28 05:17:46.436899 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Oct 28 05:17:46.441560 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Oct 28 05:17:46.441699 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Oct 28 05:17:46.457332 systemd[1]: initrd-cleanup.service: Deactivated successfully. Oct 28 05:17:46.459050 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Oct 28 05:17:46.486662 systemd[1]: sysroot-boot.mount: Deactivated successfully. Oct 28 05:17:46.565940 ignition[1116]: INFO : Ignition 2.22.0 Oct 28 05:17:46.565940 ignition[1116]: INFO : Stage: umount Oct 28 05:17:46.569423 ignition[1116]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 28 05:17:46.569423 ignition[1116]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 28 05:17:46.572311 systemd[1]: sysroot-boot.service: Deactivated successfully. Oct 28 05:17:46.572492 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Oct 28 05:17:46.578800 ignition[1116]: INFO : umount: umount passed Oct 28 05:17:46.580081 ignition[1116]: INFO : Ignition finished successfully Oct 28 05:17:46.584417 systemd[1]: ignition-mount.service: Deactivated successfully. Oct 28 05:17:46.584590 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Oct 28 05:17:46.586455 systemd[1]: Stopped target network.target - Network. Oct 28 05:17:46.591097 systemd[1]: ignition-disks.service: Deactivated successfully. Oct 28 05:17:46.591163 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Oct 28 05:17:46.594179 systemd[1]: ignition-kargs.service: Deactivated successfully. Oct 28 05:17:46.594234 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Oct 28 05:17:46.599338 systemd[1]: ignition-setup.service: Deactivated successfully. Oct 28 05:17:46.599435 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Oct 28 05:17:46.601041 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Oct 28 05:17:46.601094 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Oct 28 05:17:46.606687 systemd[1]: initrd-setup-root.service: Deactivated successfully. Oct 28 05:17:46.606746 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Oct 28 05:17:46.609314 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Oct 28 05:17:46.616772 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Oct 28 05:17:46.624976 systemd[1]: systemd-networkd.service: Deactivated successfully. Oct 28 05:17:46.625146 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Oct 28 05:17:46.633271 systemd[1]: Stopped target network-pre.target - Preparation for Network. Oct 28 05:17:46.633457 systemd[1]: systemd-networkd.socket: Deactivated successfully. Oct 28 05:17:46.633506 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Oct 28 05:17:46.639963 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Oct 28 05:17:46.644769 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Oct 28 05:17:46.646421 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 28 05:17:46.650580 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 28 05:17:46.652773 systemd[1]: systemd-resolved.service: Deactivated successfully. Oct 28 05:17:46.657571 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Oct 28 05:17:46.663460 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 28 05:17:46.663590 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 28 05:17:46.667092 systemd[1]: systemd-modules-load.service: Deactivated successfully. Oct 28 05:17:46.667158 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Oct 28 05:17:46.668570 systemd[1]: systemd-udevd.service: Deactivated successfully. Oct 28 05:17:46.668755 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 28 05:17:46.669831 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Oct 28 05:17:46.669951 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Oct 28 05:17:46.676384 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Oct 28 05:17:46.676448 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Oct 28 05:17:46.677274 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Oct 28 05:17:46.677326 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Oct 28 05:17:46.687482 systemd[1]: dracut-cmdline.service: Deactivated successfully. Oct 28 05:17:46.687562 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Oct 28 05:17:46.692452 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 28 05:17:46.692555 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 28 05:17:46.704337 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Oct 28 05:17:46.708197 systemd[1]: systemd-network-generator.service: Deactivated successfully. Oct 28 05:17:46.708289 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Oct 28 05:17:46.712328 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Oct 28 05:17:46.712419 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 28 05:17:46.714638 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Oct 28 05:17:46.714735 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 28 05:17:46.718472 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Oct 28 05:17:46.718553 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Oct 28 05:17:46.723018 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 28 05:17:46.723095 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 28 05:17:46.741329 systemd[1]: network-cleanup.service: Deactivated successfully. Oct 28 05:17:46.743045 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Oct 28 05:17:46.746431 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Oct 28 05:17:46.746565 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Oct 28 05:17:46.748530 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Oct 28 05:17:46.767611 systemd[1]: Starting initrd-switch-root.service - Switch Root... Oct 28 05:17:46.790073 systemd[1]: Switching root. Oct 28 05:17:46.836294 systemd-journald[310]: Journal stopped Oct 28 05:17:48.526687 systemd-journald[310]: Received SIGTERM from PID 1 (systemd). Oct 28 05:17:48.526755 kernel: SELinux: policy capability network_peer_controls=1 Oct 28 05:17:48.526791 kernel: SELinux: policy capability open_perms=1 Oct 28 05:17:48.526809 kernel: SELinux: policy capability extended_socket_class=1 Oct 28 05:17:48.526821 kernel: SELinux: policy capability always_check_network=0 Oct 28 05:17:48.526836 kernel: SELinux: policy capability cgroup_seclabel=1 Oct 28 05:17:48.526853 kernel: SELinux: policy capability nnp_nosuid_transition=1 Oct 28 05:17:48.526865 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Oct 28 05:17:48.526877 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Oct 28 05:17:48.526901 kernel: SELinux: policy capability userspace_initial_context=0 Oct 28 05:17:48.526914 kernel: audit: type=1403 audit(1761628667.610:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Oct 28 05:17:48.526927 systemd[1]: Successfully loaded SELinux policy in 72.331ms. Oct 28 05:17:48.526954 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 9.069ms. Oct 28 05:17:48.526968 systemd[1]: systemd 257.7 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Oct 28 05:17:48.526982 systemd[1]: Detected virtualization kvm. Oct 28 05:17:48.526994 systemd[1]: Detected architecture x86-64. Oct 28 05:17:48.527015 systemd[1]: Detected first boot. Oct 28 05:17:48.527031 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Oct 28 05:17:48.527043 zram_generator::config[1161]: No configuration found. Oct 28 05:17:48.527063 kernel: Guest personality initialized and is inactive Oct 28 05:17:48.527075 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Oct 28 05:17:48.527087 kernel: Initialized host personality Oct 28 05:17:48.527099 kernel: NET: Registered PF_VSOCK protocol family Oct 28 05:17:48.527119 systemd[1]: Populated /etc with preset unit settings. Oct 28 05:17:48.527133 systemd[1]: initrd-switch-root.service: Deactivated successfully. Oct 28 05:17:48.527146 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Oct 28 05:17:48.527168 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Oct 28 05:17:48.527182 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Oct 28 05:17:48.527195 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Oct 28 05:17:48.527212 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Oct 28 05:17:48.527233 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Oct 28 05:17:48.527247 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Oct 28 05:17:48.527260 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Oct 28 05:17:48.527273 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Oct 28 05:17:48.527287 systemd[1]: Created slice user.slice - User and Session Slice. Oct 28 05:17:48.527299 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 28 05:17:48.527321 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 28 05:17:48.527343 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Oct 28 05:17:48.527358 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Oct 28 05:17:48.527372 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Oct 28 05:17:48.527385 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 28 05:17:48.527412 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Oct 28 05:17:48.527426 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 28 05:17:48.527451 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 28 05:17:48.527466 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Oct 28 05:17:48.527480 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Oct 28 05:17:48.527495 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Oct 28 05:17:48.527509 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Oct 28 05:17:48.527522 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 28 05:17:48.527535 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 28 05:17:48.527556 systemd[1]: Reached target slices.target - Slice Units. Oct 28 05:17:48.527569 systemd[1]: Reached target swap.target - Swaps. Oct 28 05:17:48.527582 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Oct 28 05:17:48.527595 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Oct 28 05:17:48.527608 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Oct 28 05:17:48.527621 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 28 05:17:48.527634 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 28 05:17:48.527655 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 28 05:17:48.527668 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Oct 28 05:17:48.527683 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Oct 28 05:17:48.527696 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Oct 28 05:17:48.527709 systemd[1]: Mounting media.mount - External Media Directory... Oct 28 05:17:48.527722 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 28 05:17:48.527735 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Oct 28 05:17:48.527756 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Oct 28 05:17:48.527770 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Oct 28 05:17:48.527784 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Oct 28 05:17:48.527800 systemd[1]: Reached target machines.target - Containers. Oct 28 05:17:48.527813 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Oct 28 05:17:48.527827 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 28 05:17:48.527840 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 28 05:17:48.527860 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Oct 28 05:17:48.527873 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 28 05:17:48.528192 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 28 05:17:48.528208 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 28 05:17:48.528227 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Oct 28 05:17:48.528240 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 28 05:17:48.528253 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Oct 28 05:17:48.528277 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Oct 28 05:17:48.528292 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Oct 28 05:17:48.528315 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Oct 28 05:17:48.528328 systemd[1]: Stopped systemd-fsck-usr.service. Oct 28 05:17:48.528341 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 28 05:17:48.528357 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 28 05:17:48.528379 kernel: fuse: init (API version 7.41) Oct 28 05:17:48.528507 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 28 05:17:48.528523 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 28 05:17:48.528537 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Oct 28 05:17:48.528550 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Oct 28 05:17:48.528593 systemd-journald[1239]: Collecting audit messages is disabled. Oct 28 05:17:48.528617 kernel: ACPI: bus type drm_connector registered Oct 28 05:17:48.528630 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 28 05:17:48.528644 systemd-journald[1239]: Journal started Oct 28 05:17:48.528677 systemd-journald[1239]: Runtime Journal (/run/log/journal/10408c3878b84479b95d9d00a690fd2b) is 6M, max 48.1M, 42.1M free. Oct 28 05:17:48.214658 systemd[1]: Queued start job for default target multi-user.target. Oct 28 05:17:48.227649 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Oct 28 05:17:48.228216 systemd[1]: systemd-journald.service: Deactivated successfully. Oct 28 05:17:48.533565 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 28 05:17:48.537630 systemd[1]: Started systemd-journald.service - Journal Service. Oct 28 05:17:48.539842 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Oct 28 05:17:48.541770 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Oct 28 05:17:48.543743 systemd[1]: Mounted media.mount - External Media Directory. Oct 28 05:17:48.545610 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Oct 28 05:17:48.547726 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Oct 28 05:17:48.549705 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Oct 28 05:17:48.551713 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Oct 28 05:17:48.554846 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 28 05:17:48.557190 systemd[1]: modprobe@configfs.service: Deactivated successfully. Oct 28 05:17:48.557651 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Oct 28 05:17:48.559859 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 28 05:17:48.560079 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 28 05:17:48.562262 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 28 05:17:48.562654 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 28 05:17:48.564725 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 28 05:17:48.564956 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 28 05:17:48.567298 systemd[1]: modprobe@fuse.service: Deactivated successfully. Oct 28 05:17:48.567558 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Oct 28 05:17:48.569626 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 28 05:17:48.569848 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 28 05:17:48.572051 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 28 05:17:48.574364 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 28 05:17:48.577439 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Oct 28 05:17:48.579917 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Oct 28 05:17:48.597951 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 28 05:17:48.600118 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Oct 28 05:17:48.603490 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Oct 28 05:17:48.607579 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Oct 28 05:17:48.609654 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Oct 28 05:17:48.609704 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 28 05:17:48.612583 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Oct 28 05:17:48.614907 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 28 05:17:48.618208 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Oct 28 05:17:48.621110 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Oct 28 05:17:48.622984 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 28 05:17:48.624624 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Oct 28 05:17:48.626495 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 28 05:17:48.627521 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 28 05:17:48.630760 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Oct 28 05:17:48.646763 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 28 05:17:48.650237 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 28 05:17:48.653769 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Oct 28 05:17:48.656655 systemd-journald[1239]: Time spent on flushing to /var/log/journal/10408c3878b84479b95d9d00a690fd2b is 13.738ms for 1066 entries. Oct 28 05:17:48.656655 systemd-journald[1239]: System Journal (/var/log/journal/10408c3878b84479b95d9d00a690fd2b) is 8M, max 163.5M, 155.5M free. Oct 28 05:17:49.124989 systemd-journald[1239]: Received client request to flush runtime journal. Oct 28 05:17:49.125073 kernel: loop1: detected capacity change from 0 to 224512 Oct 28 05:17:49.125104 kernel: loop2: detected capacity change from 0 to 111544 Oct 28 05:17:49.125126 kernel: loop3: detected capacity change from 0 to 128912 Oct 28 05:17:49.125148 kernel: loop4: detected capacity change from 0 to 224512 Oct 28 05:17:49.125170 kernel: loop5: detected capacity change from 0 to 111544 Oct 28 05:17:49.125195 kernel: loop6: detected capacity change from 0 to 128912 Oct 28 05:17:49.125215 zram_generator::config[1323]: No configuration found. Oct 28 05:17:48.656931 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Oct 28 05:17:48.715569 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 28 05:17:48.721129 systemd-tmpfiles[1281]: ACLs are not supported, ignoring. Oct 28 05:17:48.721143 systemd-tmpfiles[1281]: ACLs are not supported, ignoring. Oct 28 05:17:48.725721 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 28 05:17:48.729212 systemd[1]: Starting systemd-sysusers.service - Create System Users... Oct 28 05:17:48.912523 (sd-merge)[1294]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw'. Oct 28 05:17:48.918425 (sd-merge)[1294]: Merged extensions into '/usr'. Oct 28 05:17:48.923187 systemd[1]: Reload requested from client PID 1280 ('systemd-sysext') (unit systemd-sysext.service)... Oct 28 05:17:48.923198 systemd[1]: Reloading... Oct 28 05:17:49.267434 systemd[1]: Reloading finished in 343 ms. Oct 28 05:17:49.306222 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Oct 28 05:17:49.308658 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Oct 28 05:17:49.311112 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Oct 28 05:17:49.313742 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Oct 28 05:17:49.316433 systemd[1]: Finished systemd-sysusers.service - Create System Users. Oct 28 05:17:49.326326 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Oct 28 05:17:49.339235 systemd[1]: Starting ensure-sysext.service... Oct 28 05:17:49.342183 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Oct 28 05:17:49.346552 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 28 05:17:49.350582 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 28 05:17:49.354000 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 28 05:17:49.368154 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Oct 28 05:17:49.370190 systemd[1]: Reload requested from client PID 1364 ('systemctl') (unit ensure-sysext.service)... Oct 28 05:17:49.370212 systemd[1]: Reloading... Oct 28 05:17:49.374144 systemd-tmpfiles[1368]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Oct 28 05:17:49.374203 systemd-tmpfiles[1368]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Oct 28 05:17:49.375021 systemd-tmpfiles[1367]: ACLs are not supported, ignoring. Oct 28 05:17:49.375146 systemd-tmpfiles[1368]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Oct 28 05:17:49.375465 systemd-tmpfiles[1368]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Oct 28 05:17:49.375605 systemd-tmpfiles[1367]: ACLs are not supported, ignoring. Oct 28 05:17:49.376511 systemd-tmpfiles[1368]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Oct 28 05:17:49.376848 systemd-tmpfiles[1368]: ACLs are not supported, ignoring. Oct 28 05:17:49.376929 systemd-tmpfiles[1368]: ACLs are not supported, ignoring. Oct 28 05:17:49.383422 systemd-tmpfiles[1368]: Detected autofs mount point /boot during canonicalization of boot. Oct 28 05:17:49.383435 systemd-tmpfiles[1368]: Skipping /boot Oct 28 05:17:49.394792 systemd-tmpfiles[1368]: Detected autofs mount point /boot during canonicalization of boot. Oct 28 05:17:49.394806 systemd-tmpfiles[1368]: Skipping /boot Oct 28 05:17:49.443422 zram_generator::config[1406]: No configuration found. Oct 28 05:17:49.506952 systemd-resolved[1366]: Positive Trust Anchors: Oct 28 05:17:49.506970 systemd-resolved[1366]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 28 05:17:49.506976 systemd-resolved[1366]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Oct 28 05:17:49.507007 systemd-resolved[1366]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 28 05:17:49.514067 systemd-resolved[1366]: Defaulting to hostname 'linux'. Oct 28 05:17:49.626450 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Oct 28 05:17:49.626901 systemd[1]: Reloading finished in 256 ms. Oct 28 05:17:49.655176 systemd[1]: Started systemd-userdbd.service - User Database Manager. Oct 28 05:17:49.657372 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 28 05:17:49.685972 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Oct 28 05:17:49.688572 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 28 05:17:49.692210 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 28 05:17:49.701566 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 28 05:17:49.705240 systemd[1]: Starting audit-rules.service - Load Audit Rules... Oct 28 05:17:49.709168 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Oct 28 05:17:49.712251 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Oct 28 05:17:49.719772 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Oct 28 05:17:49.724882 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 28 05:17:49.731842 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Oct 28 05:17:49.736880 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 28 05:17:49.738755 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 28 05:17:49.743851 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 28 05:17:49.752223 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 28 05:17:49.754313 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 28 05:17:49.754519 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 28 05:17:49.757505 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 28 05:17:49.760622 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 28 05:17:49.764134 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 28 05:17:49.764359 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 28 05:17:49.766789 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 28 05:17:49.767118 systemd-udevd[1450]: Using default interface naming scheme 'v257'. Oct 28 05:17:49.767433 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 28 05:17:49.783052 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Oct 28 05:17:49.789467 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 28 05:17:49.791634 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 28 05:17:49.795797 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 28 05:17:49.805713 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 28 05:17:49.807763 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 28 05:17:49.808008 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 28 05:17:49.809914 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Oct 28 05:17:49.812772 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 28 05:17:49.813047 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 28 05:17:49.816331 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 28 05:17:49.816604 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 28 05:17:49.821500 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 28 05:17:49.822016 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 28 05:17:49.825966 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 28 05:17:49.833611 augenrules[1481]: No rules Oct 28 05:17:49.840042 systemd[1]: audit-rules.service: Deactivated successfully. Oct 28 05:17:49.840519 systemd[1]: Finished audit-rules.service - Load Audit Rules. Oct 28 05:17:49.854893 systemd[1]: Finished ensure-sysext.service. Oct 28 05:17:49.857007 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Oct 28 05:17:49.864769 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 28 05:17:49.867534 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 28 05:17:49.871147 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 28 05:17:49.874938 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 28 05:17:49.881737 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 28 05:17:49.884465 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 28 05:17:49.884516 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 28 05:17:49.890536 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 28 05:17:49.895091 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Oct 28 05:17:49.898529 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Oct 28 05:17:49.899467 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 28 05:17:49.900503 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 28 05:17:49.920798 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 28 05:17:49.921260 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 28 05:17:49.924141 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 28 05:17:49.924390 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 28 05:17:49.926805 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 28 05:17:49.927031 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 28 05:17:49.939366 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 28 05:17:49.939485 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 28 05:17:49.941833 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Oct 28 05:17:49.978429 kernel: mousedev: PS/2 mouse device common for all mice Oct 28 05:17:50.018427 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Oct 28 05:17:50.019364 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 28 05:17:50.021908 systemd-networkd[1517]: lo: Link UP Oct 28 05:17:50.021925 systemd-networkd[1517]: lo: Gained carrier Oct 28 05:17:50.023499 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Oct 28 05:17:50.025969 systemd[1]: Reached target time-set.target - System Time Set. Oct 28 05:17:50.029668 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Oct 28 05:17:50.031593 systemd-networkd[1517]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Oct 28 05:17:50.031608 systemd-networkd[1517]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 28 05:17:50.031806 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 28 05:17:50.035719 kernel: ACPI: button: Power Button [PWRF] Oct 28 05:17:50.034518 systemd-networkd[1517]: eth0: Link UP Oct 28 05:17:50.034718 systemd[1]: Reached target network.target - Network. Oct 28 05:17:50.037306 systemd-networkd[1517]: eth0: Gained carrier Oct 28 05:17:50.037348 systemd-networkd[1517]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Oct 28 05:17:50.041755 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Oct 28 05:17:50.045533 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Oct 28 05:17:50.053468 systemd-networkd[1517]: eth0: DHCPv4 address 10.0.0.103/16, gateway 10.0.0.1 acquired from 10.0.0.1 Oct 28 05:17:50.054643 systemd-timesyncd[1518]: Network configuration changed, trying to establish connection. Oct 28 05:17:50.055896 systemd-timesyncd[1518]: Contacted time server 10.0.0.1:123 (10.0.0.1). Oct 28 05:17:50.056051 systemd-timesyncd[1518]: Initial clock synchronization to Tue 2025-10-28 05:17:49.974130 UTC. Oct 28 05:17:50.059920 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Oct 28 05:17:50.061926 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Oct 28 05:17:50.063641 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Oct 28 05:17:50.071442 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Oct 28 05:17:50.082382 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Oct 28 05:17:50.163821 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 28 05:17:50.182768 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 28 05:17:50.183336 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 28 05:17:50.190647 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 28 05:17:50.210756 kernel: kvm_amd: TSC scaling supported Oct 28 05:17:50.210843 kernel: kvm_amd: Nested Virtualization enabled Oct 28 05:17:50.210929 kernel: kvm_amd: Nested Paging enabled Oct 28 05:17:50.212975 kernel: kvm_amd: LBR virtualization supported Oct 28 05:17:50.213013 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Oct 28 05:17:50.213622 kernel: kvm_amd: Virtual GIF supported Oct 28 05:17:50.281486 kernel: EDAC MC: Ver: 3.0.0 Oct 28 05:17:50.304265 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 28 05:17:50.304289 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 28 05:17:50.317910 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 28 05:17:50.335653 ldconfig[1448]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Oct 28 05:17:50.343494 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Oct 28 05:17:50.346983 systemd[1]: Starting systemd-update-done.service - Update is Completed... Oct 28 05:17:50.368477 systemd[1]: Finished systemd-update-done.service - Update is Completed. Oct 28 05:17:50.370518 systemd[1]: Reached target sysinit.target - System Initialization. Oct 28 05:17:50.372347 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Oct 28 05:17:50.374337 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Oct 28 05:17:50.376338 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Oct 28 05:17:50.378347 systemd[1]: Started logrotate.timer - Daily rotation of log files. Oct 28 05:17:50.380176 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Oct 28 05:17:50.382168 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Oct 28 05:17:50.384163 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Oct 28 05:17:50.384194 systemd[1]: Reached target paths.target - Path Units. Oct 28 05:17:50.385649 systemd[1]: Reached target timers.target - Timer Units. Oct 28 05:17:50.388086 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Oct 28 05:17:50.391343 systemd[1]: Starting docker.socket - Docker Socket for the API... Oct 28 05:17:50.394960 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Oct 28 05:17:50.397162 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Oct 28 05:17:50.399184 systemd[1]: Reached target ssh-access.target - SSH Access Available. Oct 28 05:17:50.410033 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Oct 28 05:17:50.412338 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Oct 28 05:17:50.414788 systemd[1]: Listening on docker.socket - Docker Socket for the API. Oct 28 05:17:50.417214 systemd[1]: Reached target sockets.target - Socket Units. Oct 28 05:17:50.418785 systemd[1]: Reached target basic.target - Basic System. Oct 28 05:17:50.420334 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Oct 28 05:17:50.420366 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Oct 28 05:17:50.421432 systemd[1]: Starting containerd.service - containerd container runtime... Oct 28 05:17:50.424146 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Oct 28 05:17:50.426618 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Oct 28 05:17:50.429602 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Oct 28 05:17:50.432784 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Oct 28 05:17:50.434574 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Oct 28 05:17:50.435827 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Oct 28 05:17:50.438186 jq[1577]: false Oct 28 05:17:50.439563 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Oct 28 05:17:50.443612 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Oct 28 05:17:50.446588 google_oslogin_nss_cache[1579]: oslogin_cache_refresh[1579]: Refreshing passwd entry cache Oct 28 05:17:50.446528 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Oct 28 05:17:50.446439 oslogin_cache_refresh[1579]: Refreshing passwd entry cache Oct 28 05:17:50.450628 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Oct 28 05:17:50.452954 extend-filesystems[1578]: Found /dev/vda6 Oct 28 05:17:50.456431 extend-filesystems[1578]: Found /dev/vda9 Oct 28 05:17:50.457668 google_oslogin_nss_cache[1579]: oslogin_cache_refresh[1579]: Failure getting users, quitting Oct 28 05:17:50.457668 google_oslogin_nss_cache[1579]: oslogin_cache_refresh[1579]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Oct 28 05:17:50.457668 google_oslogin_nss_cache[1579]: oslogin_cache_refresh[1579]: Refreshing group entry cache Oct 28 05:17:50.456872 oslogin_cache_refresh[1579]: Failure getting users, quitting Oct 28 05:17:50.456896 oslogin_cache_refresh[1579]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Oct 28 05:17:50.456956 oslogin_cache_refresh[1579]: Refreshing group entry cache Oct 28 05:17:50.461915 extend-filesystems[1578]: Checking size of /dev/vda9 Oct 28 05:17:50.462524 systemd[1]: Starting systemd-logind.service - User Login Management... Oct 28 05:17:50.463359 oslogin_cache_refresh[1579]: Failure getting groups, quitting Oct 28 05:17:50.465178 google_oslogin_nss_cache[1579]: oslogin_cache_refresh[1579]: Failure getting groups, quitting Oct 28 05:17:50.465178 google_oslogin_nss_cache[1579]: oslogin_cache_refresh[1579]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Oct 28 05:17:50.465125 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Oct 28 05:17:50.463370 oslogin_cache_refresh[1579]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Oct 28 05:17:50.465679 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Oct 28 05:17:50.466409 systemd[1]: Starting update-engine.service - Update Engine... Oct 28 05:17:50.470797 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Oct 28 05:17:50.475657 extend-filesystems[1578]: Resized partition /dev/vda9 Oct 28 05:17:50.478626 extend-filesystems[1605]: resize2fs 1.47.3 (8-Jul-2025) Oct 28 05:17:50.484849 kernel: EXT4-fs (vda9): resizing filesystem from 456704 to 1784827 blocks Oct 28 05:17:50.482831 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Oct 28 05:17:50.487996 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Oct 28 05:17:50.488290 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Oct 28 05:17:50.488656 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Oct 28 05:17:50.488903 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Oct 28 05:17:50.491820 jq[1600]: true Oct 28 05:17:50.491565 systemd[1]: motdgen.service: Deactivated successfully. Oct 28 05:17:50.492153 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Oct 28 05:17:50.495541 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Oct 28 05:17:50.495841 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Oct 28 05:17:50.497496 update_engine[1598]: I20251028 05:17:50.497385 1598 main.cc:92] Flatcar Update Engine starting Oct 28 05:17:50.514980 kernel: EXT4-fs (vda9): resized filesystem to 1784827 Oct 28 05:17:50.531264 jq[1612]: true Oct 28 05:17:50.549420 extend-filesystems[1605]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Oct 28 05:17:50.549420 extend-filesystems[1605]: old_desc_blocks = 1, new_desc_blocks = 1 Oct 28 05:17:50.549420 extend-filesystems[1605]: The filesystem on /dev/vda9 is now 1784827 (4k) blocks long. Oct 28 05:17:50.553386 extend-filesystems[1578]: Resized filesystem in /dev/vda9 Oct 28 05:17:50.551955 systemd[1]: extend-filesystems.service: Deactivated successfully. Oct 28 05:17:50.552230 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Oct 28 05:17:50.566131 tar[1609]: linux-amd64/LICENSE Oct 28 05:17:50.571909 tar[1609]: linux-amd64/helm Oct 28 05:17:50.573970 dbus-daemon[1575]: [system] SELinux support is enabled Oct 28 05:17:50.574741 systemd[1]: Started dbus.service - D-Bus System Message Bus. Oct 28 05:17:50.585316 update_engine[1598]: I20251028 05:17:50.585273 1598 update_check_scheduler.cc:74] Next update check in 7m16s Oct 28 05:17:50.589642 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Oct 28 05:17:50.589675 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Oct 28 05:17:50.592796 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Oct 28 05:17:50.592816 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Oct 28 05:17:50.594974 systemd[1]: Started update-engine.service - Update Engine. Oct 28 05:17:50.599682 systemd[1]: Started locksmithd.service - Cluster reboot manager. Oct 28 05:17:50.601155 systemd-logind[1594]: Watching system buttons on /dev/input/event2 (Power Button) Oct 28 05:17:50.602461 systemd-logind[1594]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Oct 28 05:17:50.603035 systemd-logind[1594]: New seat seat0. Oct 28 05:17:50.606345 systemd[1]: Started systemd-logind.service - User Login Management. Oct 28 05:17:50.612556 bash[1645]: Updated "/home/core/.ssh/authorized_keys" Oct 28 05:17:50.611283 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Oct 28 05:17:50.615673 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Oct 28 05:17:50.686221 locksmithd[1646]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Oct 28 05:17:50.786910 containerd[1617]: time="2025-10-28T05:17:50Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Oct 28 05:17:50.788858 containerd[1617]: time="2025-10-28T05:17:50.788820795Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Oct 28 05:17:50.803273 containerd[1617]: time="2025-10-28T05:17:50.803192838Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="12.995µs" Oct 28 05:17:50.804416 containerd[1617]: time="2025-10-28T05:17:50.803450912Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Oct 28 05:17:50.804416 containerd[1617]: time="2025-10-28T05:17:50.803492039Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Oct 28 05:17:50.804416 containerd[1617]: time="2025-10-28T05:17:50.803824703Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Oct 28 05:17:50.804416 containerd[1617]: time="2025-10-28T05:17:50.803848378Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Oct 28 05:17:50.804416 containerd[1617]: time="2025-10-28T05:17:50.803888623Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Oct 28 05:17:50.804416 containerd[1617]: time="2025-10-28T05:17:50.804011994Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Oct 28 05:17:50.804416 containerd[1617]: time="2025-10-28T05:17:50.804036651Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Oct 28 05:17:50.804899 containerd[1617]: time="2025-10-28T05:17:50.804858652Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Oct 28 05:17:50.805006 containerd[1617]: time="2025-10-28T05:17:50.804974770Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Oct 28 05:17:50.805289 containerd[1617]: time="2025-10-28T05:17:50.805201996Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Oct 28 05:17:50.805289 containerd[1617]: time="2025-10-28T05:17:50.805288849Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Oct 28 05:17:50.805601 containerd[1617]: time="2025-10-28T05:17:50.805559026Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Oct 28 05:17:50.806055 containerd[1617]: time="2025-10-28T05:17:50.806019439Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Oct 28 05:17:50.806107 containerd[1617]: time="2025-10-28T05:17:50.806069232Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Oct 28 05:17:50.806107 containerd[1617]: time="2025-10-28T05:17:50.806083609Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Oct 28 05:17:50.806191 containerd[1617]: time="2025-10-28T05:17:50.806127652Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Oct 28 05:17:50.806939 containerd[1617]: time="2025-10-28T05:17:50.806887737Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Oct 28 05:17:50.807011 containerd[1617]: time="2025-10-28T05:17:50.806982385Z" level=info msg="metadata content store policy set" policy=shared Oct 28 05:17:50.813344 containerd[1617]: time="2025-10-28T05:17:50.813273583Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Oct 28 05:17:50.813344 containerd[1617]: time="2025-10-28T05:17:50.813334207Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Oct 28 05:17:50.813344 containerd[1617]: time="2025-10-28T05:17:50.813349536Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813361508Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813375635Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813387056Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813412003Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813423885Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813448431Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813460293Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813469460Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Oct 28 05:17:50.813572 containerd[1617]: time="2025-10-28T05:17:50.813481703Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813645841Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813665608Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813678803Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813690455Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813707617Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813724970Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813736341Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813746821Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813767509Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813779722Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Oct 28 05:17:50.813822 containerd[1617]: time="2025-10-28T05:17:50.813789240Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Oct 28 05:17:50.814281 containerd[1617]: time="2025-10-28T05:17:50.813857408Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Oct 28 05:17:50.814281 containerd[1617]: time="2025-10-28T05:17:50.813913042Z" level=info msg="Start snapshots syncer" Oct 28 05:17:50.814281 containerd[1617]: time="2025-10-28T05:17:50.813932659Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Oct 28 05:17:50.814455 containerd[1617]: time="2025-10-28T05:17:50.814165486Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Oct 28 05:17:50.814455 containerd[1617]: time="2025-10-28T05:17:50.814224657Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Oct 28 05:17:50.815902 containerd[1617]: time="2025-10-28T05:17:50.815872657Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Oct 28 05:17:50.816041 containerd[1617]: time="2025-10-28T05:17:50.816012059Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Oct 28 05:17:50.816041 containerd[1617]: time="2025-10-28T05:17:50.816042586Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Oct 28 05:17:50.816138 containerd[1617]: time="2025-10-28T05:17:50.816060720Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Oct 28 05:17:50.816138 containerd[1617]: time="2025-10-28T05:17:50.816072652Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Oct 28 05:17:50.816138 containerd[1617]: time="2025-10-28T05:17:50.816083653Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Oct 28 05:17:50.816138 containerd[1617]: time="2025-10-28T05:17:50.816093301Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Oct 28 05:17:50.816138 containerd[1617]: time="2025-10-28T05:17:50.816103340Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Oct 28 05:17:50.816138 containerd[1617]: time="2025-10-28T05:17:50.816125041Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816140620Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816164084Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816217163Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816231841Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816320708Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816335125Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816343791Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816352748Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816362436Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816374318Z" level=info msg="runtime interface created" Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816380089Z" level=info msg="created NRI interface" Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816413952Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Oct 28 05:17:50.816425 containerd[1617]: time="2025-10-28T05:17:50.816429632Z" level=info msg="Connect containerd service" Oct 28 05:17:50.816970 containerd[1617]: time="2025-10-28T05:17:50.816457614Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Oct 28 05:17:50.817557 containerd[1617]: time="2025-10-28T05:17:50.817509347Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 28 05:17:50.876101 tar[1609]: linux-amd64/README.md Oct 28 05:17:50.884144 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Oct 28 05:17:50.913648 containerd[1617]: time="2025-10-28T05:17:50.913583179Z" level=info msg="Start subscribing containerd event" Oct 28 05:17:50.913789 containerd[1617]: time="2025-10-28T05:17:50.913651217Z" level=info msg="Start recovering state" Oct 28 05:17:50.913789 containerd[1617]: time="2025-10-28T05:17:50.913783605Z" level=info msg="Start event monitor" Oct 28 05:17:50.913830 containerd[1617]: time="2025-10-28T05:17:50.913791169Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Oct 28 05:17:50.913873 containerd[1617]: time="2025-10-28T05:17:50.913800046Z" level=info msg="Start cni network conf syncer for default" Oct 28 05:17:50.913898 containerd[1617]: time="2025-10-28T05:17:50.913880567Z" level=info msg=serving... address=/run/containerd/containerd.sock Oct 28 05:17:50.913947 containerd[1617]: time="2025-10-28T05:17:50.913933126Z" level=info msg="Start streaming server" Oct 28 05:17:50.913996 containerd[1617]: time="2025-10-28T05:17:50.913978791Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Oct 28 05:17:50.913996 containerd[1617]: time="2025-10-28T05:17:50.913991475Z" level=info msg="runtime interface starting up..." Oct 28 05:17:50.914143 containerd[1617]: time="2025-10-28T05:17:50.913998037Z" level=info msg="starting plugins..." Oct 28 05:17:50.914143 containerd[1617]: time="2025-10-28T05:17:50.914021281Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Oct 28 05:17:50.914888 containerd[1617]: time="2025-10-28T05:17:50.914866125Z" level=info msg="containerd successfully booted in 0.128900s" Oct 28 05:17:50.914970 systemd[1]: Started containerd.service - containerd container runtime. Oct 28 05:17:50.959980 sshd_keygen[1601]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Oct 28 05:17:50.985084 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Oct 28 05:17:50.988806 systemd[1]: Starting issuegen.service - Generate /run/issue... Oct 28 05:17:51.016855 systemd[1]: issuegen.service: Deactivated successfully. Oct 28 05:17:51.017199 systemd[1]: Finished issuegen.service - Generate /run/issue. Oct 28 05:17:51.020682 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Oct 28 05:17:51.057626 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Oct 28 05:17:51.061326 systemd[1]: Started getty@tty1.service - Getty on tty1. Oct 28 05:17:51.064362 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Oct 28 05:17:51.066312 systemd[1]: Reached target getty.target - Login Prompts. Oct 28 05:17:51.440629 systemd-networkd[1517]: eth0: Gained IPv6LL Oct 28 05:17:51.444274 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Oct 28 05:17:51.447038 systemd[1]: Reached target network-online.target - Network is Online. Oct 28 05:17:51.450411 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Oct 28 05:17:51.453506 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 28 05:17:51.456446 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Oct 28 05:17:51.496652 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Oct 28 05:17:51.501050 systemd[1]: coreos-metadata.service: Deactivated successfully. Oct 28 05:17:51.501329 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Oct 28 05:17:51.503632 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Oct 28 05:17:52.954358 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Oct 28 05:17:52.957835 systemd[1]: Started sshd@0-10.0.0.103:22-10.0.0.1:36740.service - OpenSSH per-connection server daemon (10.0.0.1:36740). Oct 28 05:17:53.233772 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 28 05:17:53.236602 systemd[1]: Reached target multi-user.target - Multi-User System. Oct 28 05:17:53.238800 systemd[1]: Startup finished in 3.098s (kernel) + 7.373s (initrd) + 5.698s (userspace) = 16.170s. Oct 28 05:17:53.246882 (kubelet)[1721]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 28 05:17:53.256789 sshd[1713]: Accepted publickey for core from 10.0.0.1 port 36740 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:17:53.261005 sshd-session[1713]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:17:53.270094 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Oct 28 05:17:53.271476 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Oct 28 05:17:53.278216 systemd-logind[1594]: New session 1 of user core. Oct 28 05:17:53.299564 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Oct 28 05:17:53.303009 systemd[1]: Starting user@500.service - User Manager for UID 500... Oct 28 05:17:53.317138 (systemd)[1726]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Oct 28 05:17:53.319791 systemd-logind[1594]: New session c1 of user core. Oct 28 05:17:53.465361 systemd[1726]: Queued start job for default target default.target. Oct 28 05:17:53.484818 systemd[1726]: Created slice app.slice - User Application Slice. Oct 28 05:17:53.484852 systemd[1726]: Reached target paths.target - Paths. Oct 28 05:17:53.484907 systemd[1726]: Reached target timers.target - Timers. Oct 28 05:17:53.486945 systemd[1726]: Starting dbus.socket - D-Bus User Message Bus Socket... Oct 28 05:17:53.500324 systemd[1726]: Listening on dbus.socket - D-Bus User Message Bus Socket. Oct 28 05:17:53.500486 systemd[1726]: Reached target sockets.target - Sockets. Oct 28 05:17:53.500526 systemd[1726]: Reached target basic.target - Basic System. Oct 28 05:17:53.500573 systemd[1726]: Reached target default.target - Main User Target. Oct 28 05:17:53.500606 systemd[1726]: Startup finished in 170ms. Oct 28 05:17:53.501515 systemd[1]: Started user@500.service - User Manager for UID 500. Oct 28 05:17:53.503905 systemd[1]: Started session-1.scope - Session 1 of User core. Oct 28 05:17:53.524418 systemd[1]: Started sshd@1-10.0.0.103:22-10.0.0.1:36748.service - OpenSSH per-connection server daemon (10.0.0.1:36748). Oct 28 05:17:53.576522 sshd[1743]: Accepted publickey for core from 10.0.0.1 port 36748 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:17:53.578211 sshd-session[1743]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:17:53.584208 systemd-logind[1594]: New session 2 of user core. Oct 28 05:17:53.592454 systemd[1]: Started session-2.scope - Session 2 of User core. Oct 28 05:17:53.617921 sshd[1747]: Connection closed by 10.0.0.1 port 36748 Oct 28 05:17:53.618272 sshd-session[1743]: pam_unix(sshd:session): session closed for user core Oct 28 05:17:53.630984 systemd[1]: sshd@1-10.0.0.103:22-10.0.0.1:36748.service: Deactivated successfully. Oct 28 05:17:53.633434 systemd[1]: session-2.scope: Deactivated successfully. Oct 28 05:17:53.634267 systemd-logind[1594]: Session 2 logged out. Waiting for processes to exit. Oct 28 05:17:53.637815 systemd[1]: Started sshd@2-10.0.0.103:22-10.0.0.1:36760.service - OpenSSH per-connection server daemon (10.0.0.1:36760). Oct 28 05:17:53.638931 systemd-logind[1594]: Removed session 2. Oct 28 05:17:53.800187 sshd[1754]: Accepted publickey for core from 10.0.0.1 port 36760 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:17:53.801740 sshd-session[1754]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:17:53.807437 systemd-logind[1594]: New session 3 of user core. Oct 28 05:17:53.814588 systemd[1]: Started session-3.scope - Session 3 of User core. Oct 28 05:17:53.825003 sshd[1757]: Connection closed by 10.0.0.1 port 36760 Oct 28 05:17:53.825368 sshd-session[1754]: pam_unix(sshd:session): session closed for user core Oct 28 05:17:53.837468 systemd[1]: sshd@2-10.0.0.103:22-10.0.0.1:36760.service: Deactivated successfully. Oct 28 05:17:53.839594 systemd[1]: session-3.scope: Deactivated successfully. Oct 28 05:17:53.840406 systemd-logind[1594]: Session 3 logged out. Waiting for processes to exit. Oct 28 05:17:53.843939 systemd[1]: Started sshd@3-10.0.0.103:22-10.0.0.1:36772.service - OpenSSH per-connection server daemon (10.0.0.1:36772). Oct 28 05:17:53.845523 systemd-logind[1594]: Removed session 3. Oct 28 05:17:54.161044 kubelet[1721]: E1028 05:17:54.160892 1721 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 28 05:17:54.165199 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 28 05:17:54.165461 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 28 05:17:54.167107 systemd[1]: kubelet.service: Consumed 2.509s CPU time, 266.8M memory peak. Oct 28 05:17:54.203588 sshd[1763]: Accepted publickey for core from 10.0.0.1 port 36772 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:17:54.205070 sshd-session[1763]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:17:54.209609 systemd-logind[1594]: New session 4 of user core. Oct 28 05:17:54.227529 systemd[1]: Started session-4.scope - Session 4 of User core. Oct 28 05:17:54.240554 sshd[1767]: Connection closed by 10.0.0.1 port 36772 Oct 28 05:17:54.240895 sshd-session[1763]: pam_unix(sshd:session): session closed for user core Oct 28 05:17:54.250059 systemd[1]: sshd@3-10.0.0.103:22-10.0.0.1:36772.service: Deactivated successfully. Oct 28 05:17:54.252063 systemd[1]: session-4.scope: Deactivated successfully. Oct 28 05:17:54.252791 systemd-logind[1594]: Session 4 logged out. Waiting for processes to exit. Oct 28 05:17:54.255554 systemd[1]: Started sshd@4-10.0.0.103:22-10.0.0.1:36774.service - OpenSSH per-connection server daemon (10.0.0.1:36774). Oct 28 05:17:54.256109 systemd-logind[1594]: Removed session 4. Oct 28 05:17:54.306856 sshd[1773]: Accepted publickey for core from 10.0.0.1 port 36774 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:17:54.308124 sshd-session[1773]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:17:54.312440 systemd-logind[1594]: New session 5 of user core. Oct 28 05:17:54.319522 systemd[1]: Started session-5.scope - Session 5 of User core. Oct 28 05:17:54.340007 sudo[1777]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Oct 28 05:17:54.340315 sudo[1777]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 28 05:17:54.359258 sudo[1777]: pam_unix(sudo:session): session closed for user root Oct 28 05:17:54.360848 sshd[1776]: Connection closed by 10.0.0.1 port 36774 Oct 28 05:17:54.361165 sshd-session[1773]: pam_unix(sshd:session): session closed for user core Oct 28 05:17:54.374095 systemd[1]: sshd@4-10.0.0.103:22-10.0.0.1:36774.service: Deactivated successfully. Oct 28 05:17:54.376059 systemd[1]: session-5.scope: Deactivated successfully. Oct 28 05:17:54.376866 systemd-logind[1594]: Session 5 logged out. Waiting for processes to exit. Oct 28 05:17:54.379905 systemd[1]: Started sshd@5-10.0.0.103:22-10.0.0.1:36790.service - OpenSSH per-connection server daemon (10.0.0.1:36790). Oct 28 05:17:54.380573 systemd-logind[1594]: Removed session 5. Oct 28 05:17:54.435476 sshd[1783]: Accepted publickey for core from 10.0.0.1 port 36790 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:17:54.436905 sshd-session[1783]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:17:54.441448 systemd-logind[1594]: New session 6 of user core. Oct 28 05:17:54.455536 systemd[1]: Started session-6.scope - Session 6 of User core. Oct 28 05:17:54.469364 sudo[1788]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Oct 28 05:17:54.469673 sudo[1788]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 28 05:17:54.476177 sudo[1788]: pam_unix(sudo:session): session closed for user root Oct 28 05:17:54.484017 sudo[1787]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Oct 28 05:17:54.484334 sudo[1787]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 28 05:17:54.494510 systemd[1]: Starting audit-rules.service - Load Audit Rules... Oct 28 05:17:54.549968 augenrules[1810]: No rules Oct 28 05:17:54.551624 systemd[1]: audit-rules.service: Deactivated successfully. Oct 28 05:17:54.551899 systemd[1]: Finished audit-rules.service - Load Audit Rules. Oct 28 05:17:54.553072 sudo[1787]: pam_unix(sudo:session): session closed for user root Oct 28 05:17:54.554821 sshd[1786]: Connection closed by 10.0.0.1 port 36790 Oct 28 05:17:54.555132 sshd-session[1783]: pam_unix(sshd:session): session closed for user core Oct 28 05:17:54.568689 systemd[1]: sshd@5-10.0.0.103:22-10.0.0.1:36790.service: Deactivated successfully. Oct 28 05:17:54.570213 systemd[1]: session-6.scope: Deactivated successfully. Oct 28 05:17:54.570938 systemd-logind[1594]: Session 6 logged out. Waiting for processes to exit. Oct 28 05:17:54.573352 systemd[1]: Started sshd@6-10.0.0.103:22-10.0.0.1:36806.service - OpenSSH per-connection server daemon (10.0.0.1:36806). Oct 28 05:17:54.573953 systemd-logind[1594]: Removed session 6. Oct 28 05:17:54.635561 sshd[1819]: Accepted publickey for core from 10.0.0.1 port 36806 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:17:54.637220 sshd-session[1819]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:17:54.641357 systemd-logind[1594]: New session 7 of user core. Oct 28 05:17:54.656529 systemd[1]: Started session-7.scope - Session 7 of User core. Oct 28 05:17:54.670042 sudo[1823]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Oct 28 05:17:54.670347 sudo[1823]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 28 05:17:55.440936 systemd[1]: Starting docker.service - Docker Application Container Engine... Oct 28 05:17:55.471795 (dockerd)[1844]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Oct 28 05:17:56.032965 dockerd[1844]: time="2025-10-28T05:17:56.032884450Z" level=info msg="Starting up" Oct 28 05:17:56.033854 dockerd[1844]: time="2025-10-28T05:17:56.033805115Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Oct 28 05:17:56.053055 dockerd[1844]: time="2025-10-28T05:17:56.053001350Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Oct 28 05:17:56.570614 dockerd[1844]: time="2025-10-28T05:17:56.570548610Z" level=info msg="Loading containers: start." Oct 28 05:17:56.581434 kernel: Initializing XFRM netlink socket Oct 28 05:17:56.889718 systemd-networkd[1517]: docker0: Link UP Oct 28 05:17:56.895210 dockerd[1844]: time="2025-10-28T05:17:56.895155105Z" level=info msg="Loading containers: done." Oct 28 05:17:56.910062 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2994636974-merged.mount: Deactivated successfully. Oct 28 05:17:56.912546 dockerd[1844]: time="2025-10-28T05:17:56.912494407Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Oct 28 05:17:56.912658 dockerd[1844]: time="2025-10-28T05:17:56.912601025Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Oct 28 05:17:56.912757 dockerd[1844]: time="2025-10-28T05:17:56.912719730Z" level=info msg="Initializing buildkit" Oct 28 05:17:56.942195 dockerd[1844]: time="2025-10-28T05:17:56.942153311Z" level=info msg="Completed buildkit initialization" Oct 28 05:17:56.948353 dockerd[1844]: time="2025-10-28T05:17:56.948315388Z" level=info msg="Daemon has completed initialization" Oct 28 05:17:56.948471 dockerd[1844]: time="2025-10-28T05:17:56.948419050Z" level=info msg="API listen on /run/docker.sock" Oct 28 05:17:56.948675 systemd[1]: Started docker.service - Docker Application Container Engine. Oct 28 05:17:57.827548 containerd[1617]: time="2025-10-28T05:17:57.827491063Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Oct 28 05:17:58.550097 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount762836476.mount: Deactivated successfully. Oct 28 05:17:59.929044 containerd[1617]: time="2025-10-28T05:17:59.928955919Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:17:59.929672 containerd[1617]: time="2025-10-28T05:17:59.929609462Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837916" Oct 28 05:17:59.930873 containerd[1617]: time="2025-10-28T05:17:59.930831797Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:17:59.933570 containerd[1617]: time="2025-10-28T05:17:59.933511983Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:17:59.934481 containerd[1617]: time="2025-10-28T05:17:59.934410992Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 2.106830314s" Oct 28 05:17:59.934481 containerd[1617]: time="2025-10-28T05:17:59.934481441Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Oct 28 05:17:59.935741 containerd[1617]: time="2025-10-28T05:17:59.935712755Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Oct 28 05:18:01.070985 containerd[1617]: time="2025-10-28T05:18:01.070636634Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787027" Oct 28 05:18:01.070985 containerd[1617]: time="2025-10-28T05:18:01.070806471Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:01.072175 containerd[1617]: time="2025-10-28T05:18:01.072123670Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:01.078602 containerd[1617]: time="2025-10-28T05:18:01.078565453Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:01.079662 containerd[1617]: time="2025-10-28T05:18:01.079619091Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 1.143875359s" Oct 28 05:18:01.079662 containerd[1617]: time="2025-10-28T05:18:01.079658081Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Oct 28 05:18:01.080191 containerd[1617]: time="2025-10-28T05:18:01.080150572Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Oct 28 05:18:02.333947 containerd[1617]: time="2025-10-28T05:18:02.333881240Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:02.334696 containerd[1617]: time="2025-10-28T05:18:02.334670238Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176289" Oct 28 05:18:02.335956 containerd[1617]: time="2025-10-28T05:18:02.335913554Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:02.338154 containerd[1617]: time="2025-10-28T05:18:02.338121889Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:02.339067 containerd[1617]: time="2025-10-28T05:18:02.339024077Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 1.258837881s" Oct 28 05:18:02.339067 containerd[1617]: time="2025-10-28T05:18:02.339054848Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Oct 28 05:18:02.339734 containerd[1617]: time="2025-10-28T05:18:02.339697548Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Oct 28 05:18:03.364648 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount275203488.mount: Deactivated successfully. Oct 28 05:18:04.253686 containerd[1617]: time="2025-10-28T05:18:04.253626726Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:04.254712 containerd[1617]: time="2025-10-28T05:18:04.254688337Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924206" Oct 28 05:18:04.255887 containerd[1617]: time="2025-10-28T05:18:04.255851945Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:04.257861 containerd[1617]: time="2025-10-28T05:18:04.257808620Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:04.258249 containerd[1617]: time="2025-10-28T05:18:04.258203354Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 1.918476823s" Oct 28 05:18:04.258249 containerd[1617]: time="2025-10-28T05:18:04.258230555Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Oct 28 05:18:04.258766 containerd[1617]: time="2025-10-28T05:18:04.258730276Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Oct 28 05:18:04.416053 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Oct 28 05:18:04.418028 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 28 05:18:04.652058 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 28 05:18:04.656352 (kubelet)[2144]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 28 05:18:05.516540 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount362418054.mount: Deactivated successfully. Oct 28 05:18:05.527139 kubelet[2144]: E1028 05:18:05.527050 2144 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 28 05:18:05.541601 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 28 05:18:05.541854 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 28 05:18:05.542369 systemd[1]: kubelet.service: Consumed 744ms CPU time, 111.3M memory peak. Oct 28 05:18:06.394895 containerd[1617]: time="2025-10-28T05:18:06.394831755Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:06.395630 containerd[1617]: time="2025-10-28T05:18:06.395577035Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Oct 28 05:18:06.396756 containerd[1617]: time="2025-10-28T05:18:06.396719415Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:06.399302 containerd[1617]: time="2025-10-28T05:18:06.399242250Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:06.400279 containerd[1617]: time="2025-10-28T05:18:06.400242097Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 2.141482489s" Oct 28 05:18:06.400279 containerd[1617]: time="2025-10-28T05:18:06.400277425Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Oct 28 05:18:06.401313 containerd[1617]: time="2025-10-28T05:18:06.401290131Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Oct 28 05:18:06.991737 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4112903184.mount: Deactivated successfully. Oct 28 05:18:06.998964 containerd[1617]: time="2025-10-28T05:18:06.998916166Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 28 05:18:06.999771 containerd[1617]: time="2025-10-28T05:18:06.999713364Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Oct 28 05:18:07.000921 containerd[1617]: time="2025-10-28T05:18:07.000883711Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 28 05:18:07.003679 containerd[1617]: time="2025-10-28T05:18:07.003630635Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 28 05:18:07.004109 containerd[1617]: time="2025-10-28T05:18:07.004070224Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 602.755044ms" Oct 28 05:18:07.004109 containerd[1617]: time="2025-10-28T05:18:07.004103837Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Oct 28 05:18:07.004675 containerd[1617]: time="2025-10-28T05:18:07.004650833Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Oct 28 05:18:07.622013 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount390227910.mount: Deactivated successfully. Oct 28 05:18:09.740431 containerd[1617]: time="2025-10-28T05:18:09.740350132Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:09.741021 containerd[1617]: time="2025-10-28T05:18:09.740971990Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Oct 28 05:18:09.742429 containerd[1617]: time="2025-10-28T05:18:09.742382325Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:09.745189 containerd[1617]: time="2025-10-28T05:18:09.745140996Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:09.746122 containerd[1617]: time="2025-10-28T05:18:09.746077605Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.741396708s" Oct 28 05:18:09.746122 containerd[1617]: time="2025-10-28T05:18:09.746119737Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Oct 28 05:18:11.749506 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 28 05:18:11.749682 systemd[1]: kubelet.service: Consumed 744ms CPU time, 111.3M memory peak. Oct 28 05:18:11.751976 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 28 05:18:11.782762 systemd[1]: Reload requested from client PID 2294 ('systemctl') (unit session-7.scope)... Oct 28 05:18:11.782799 systemd[1]: Reloading... Oct 28 05:18:11.910514 zram_generator::config[2338]: No configuration found. Oct 28 05:18:12.276272 systemd[1]: Reloading finished in 492 ms. Oct 28 05:18:12.344300 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Oct 28 05:18:12.344424 systemd[1]: kubelet.service: Failed with result 'signal'. Oct 28 05:18:12.344759 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 28 05:18:12.344813 systemd[1]: kubelet.service: Consumed 169ms CPU time, 98.3M memory peak. Oct 28 05:18:12.346483 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 28 05:18:12.544765 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 28 05:18:12.562728 (kubelet)[2386]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 28 05:18:12.603555 kubelet[2386]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 28 05:18:12.603555 kubelet[2386]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 28 05:18:12.603555 kubelet[2386]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 28 05:18:12.603872 kubelet[2386]: I1028 05:18:12.603645 2386 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 28 05:18:13.107071 kubelet[2386]: I1028 05:18:13.106996 2386 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 28 05:18:13.107071 kubelet[2386]: I1028 05:18:13.107043 2386 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 28 05:18:13.107380 kubelet[2386]: I1028 05:18:13.107345 2386 server.go:954] "Client rotation is on, will bootstrap in background" Oct 28 05:18:13.127768 kubelet[2386]: I1028 05:18:13.127705 2386 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 28 05:18:13.131191 kubelet[2386]: E1028 05:18:13.130151 2386 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.103:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:13.136468 kubelet[2386]: I1028 05:18:13.136432 2386 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 28 05:18:13.142318 kubelet[2386]: I1028 05:18:13.142275 2386 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 28 05:18:13.143255 kubelet[2386]: I1028 05:18:13.143219 2386 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 28 05:18:13.143502 kubelet[2386]: I1028 05:18:13.143254 2386 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 28 05:18:13.143692 kubelet[2386]: I1028 05:18:13.143503 2386 topology_manager.go:138] "Creating topology manager with none policy" Oct 28 05:18:13.143692 kubelet[2386]: I1028 05:18:13.143513 2386 container_manager_linux.go:304] "Creating device plugin manager" Oct 28 05:18:13.143692 kubelet[2386]: I1028 05:18:13.143676 2386 state_mem.go:36] "Initialized new in-memory state store" Oct 28 05:18:13.146557 kubelet[2386]: I1028 05:18:13.146526 2386 kubelet.go:446] "Attempting to sync node with API server" Oct 28 05:18:13.146627 kubelet[2386]: I1028 05:18:13.146566 2386 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 28 05:18:13.146627 kubelet[2386]: I1028 05:18:13.146593 2386 kubelet.go:352] "Adding apiserver pod source" Oct 28 05:18:13.146627 kubelet[2386]: I1028 05:18:13.146609 2386 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 28 05:18:13.149639 kubelet[2386]: I1028 05:18:13.148903 2386 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Oct 28 05:18:13.149639 kubelet[2386]: I1028 05:18:13.149439 2386 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 28 05:18:13.150121 kubelet[2386]: W1028 05:18:13.150088 2386 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Oct 28 05:18:13.150337 kubelet[2386]: W1028 05:18:13.150285 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.103:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:13.150386 kubelet[2386]: E1028 05:18:13.150347 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.103:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:13.150436 kubelet[2386]: W1028 05:18:13.150390 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.103:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:13.150467 kubelet[2386]: E1028 05:18:13.150442 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.103:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:13.151967 kubelet[2386]: I1028 05:18:13.151937 2386 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 28 05:18:13.152020 kubelet[2386]: I1028 05:18:13.151975 2386 server.go:1287] "Started kubelet" Oct 28 05:18:13.153303 kubelet[2386]: I1028 05:18:13.153255 2386 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 28 05:18:13.153380 kubelet[2386]: I1028 05:18:13.153362 2386 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 28 05:18:13.153725 kubelet[2386]: I1028 05:18:13.153680 2386 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 28 05:18:13.153858 kubelet[2386]: I1028 05:18:13.153755 2386 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 28 05:18:13.154902 kubelet[2386]: I1028 05:18:13.154872 2386 server.go:479] "Adding debug handlers to kubelet server" Oct 28 05:18:13.157531 kubelet[2386]: I1028 05:18:13.156983 2386 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 28 05:18:13.158121 kubelet[2386]: E1028 05:18:13.158101 2386 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 28 05:18:13.158496 kubelet[2386]: E1028 05:18:13.158469 2386 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 28 05:18:13.158544 kubelet[2386]: I1028 05:18:13.158510 2386 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 28 05:18:13.158705 kubelet[2386]: I1028 05:18:13.158688 2386 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 28 05:18:13.158753 kubelet[2386]: I1028 05:18:13.158744 2386 reconciler.go:26] "Reconciler: start to sync state" Oct 28 05:18:13.159141 kubelet[2386]: W1028 05:18:13.159094 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.103:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:13.159189 kubelet[2386]: E1028 05:18:13.159146 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.103:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:13.160485 kubelet[2386]: E1028 05:18:13.159627 2386 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.103:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.103:6443: connect: connection refused" interval="200ms" Oct 28 05:18:13.160485 kubelet[2386]: I1028 05:18:13.159796 2386 factory.go:221] Registration of the systemd container factory successfully Oct 28 05:18:13.160485 kubelet[2386]: I1028 05:18:13.159886 2386 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 28 05:18:13.161139 kubelet[2386]: I1028 05:18:13.161123 2386 factory.go:221] Registration of the containerd container factory successfully Oct 28 05:18:13.162025 kubelet[2386]: E1028 05:18:13.160799 2386 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.103:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.103:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18728ff681753b9d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-10-28 05:18:13.151955869 +0000 UTC m=+0.584694219,LastTimestamp:2025-10-28 05:18:13.151955869 +0000 UTC m=+0.584694219,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Oct 28 05:18:13.180892 kubelet[2386]: I1028 05:18:13.180748 2386 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 28 05:18:13.181864 kubelet[2386]: I1028 05:18:13.181843 2386 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 28 05:18:13.182062 kubelet[2386]: I1028 05:18:13.182045 2386 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 28 05:18:13.182140 kubelet[2386]: I1028 05:18:13.182130 2386 state_mem.go:36] "Initialized new in-memory state store" Oct 28 05:18:13.183683 kubelet[2386]: I1028 05:18:13.183647 2386 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 28 05:18:13.183683 kubelet[2386]: I1028 05:18:13.183681 2386 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 28 05:18:13.183989 kubelet[2386]: I1028 05:18:13.183708 2386 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 28 05:18:13.183989 kubelet[2386]: I1028 05:18:13.183718 2386 kubelet.go:2382] "Starting kubelet main sync loop" Oct 28 05:18:13.183989 kubelet[2386]: E1028 05:18:13.183778 2386 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 28 05:18:13.184559 kubelet[2386]: W1028 05:18:13.184530 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.103:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:13.184606 kubelet[2386]: E1028 05:18:13.184578 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.103:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:13.259454 kubelet[2386]: E1028 05:18:13.259371 2386 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 28 05:18:13.284735 kubelet[2386]: E1028 05:18:13.284640 2386 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 28 05:18:13.359619 kubelet[2386]: E1028 05:18:13.359482 2386 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 28 05:18:13.361077 kubelet[2386]: E1028 05:18:13.361046 2386 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.103:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.103:6443: connect: connection refused" interval="400ms" Oct 28 05:18:13.460430 kubelet[2386]: E1028 05:18:13.460361 2386 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 28 05:18:13.485690 kubelet[2386]: E1028 05:18:13.485610 2386 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 28 05:18:13.560481 kubelet[2386]: E1028 05:18:13.560436 2386 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 28 05:18:13.567523 kubelet[2386]: I1028 05:18:13.567505 2386 policy_none.go:49] "None policy: Start" Oct 28 05:18:13.567523 kubelet[2386]: I1028 05:18:13.567524 2386 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 28 05:18:13.567626 kubelet[2386]: I1028 05:18:13.567537 2386 state_mem.go:35] "Initializing new in-memory state store" Oct 28 05:18:13.575884 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Oct 28 05:18:13.605838 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Oct 28 05:18:13.609144 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Oct 28 05:18:13.617381 kubelet[2386]: I1028 05:18:13.617284 2386 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 28 05:18:13.617714 kubelet[2386]: I1028 05:18:13.617557 2386 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 28 05:18:13.617714 kubelet[2386]: I1028 05:18:13.617569 2386 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 28 05:18:13.618166 kubelet[2386]: I1028 05:18:13.617943 2386 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 28 05:18:13.618865 kubelet[2386]: E1028 05:18:13.618836 2386 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 28 05:18:13.618865 kubelet[2386]: E1028 05:18:13.618869 2386 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Oct 28 05:18:13.719718 kubelet[2386]: I1028 05:18:13.719664 2386 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Oct 28 05:18:13.720143 kubelet[2386]: E1028 05:18:13.720100 2386 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.103:6443/api/v1/nodes\": dial tcp 10.0.0.103:6443: connect: connection refused" node="localhost" Oct 28 05:18:13.761919 kubelet[2386]: E1028 05:18:13.761874 2386 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.103:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.103:6443: connect: connection refused" interval="800ms" Oct 28 05:18:13.893792 systemd[1]: Created slice kubepods-burstable-pod2abbad858f7a7505c8a8b4aa1ce70949.slice - libcontainer container kubepods-burstable-pod2abbad858f7a7505c8a8b4aa1ce70949.slice. Oct 28 05:18:13.915865 kubelet[2386]: E1028 05:18:13.915842 2386 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Oct 28 05:18:13.919246 systemd[1]: Created slice kubepods-burstable-pod4654b122dbb389158fe3c0766e603624.slice - libcontainer container kubepods-burstable-pod4654b122dbb389158fe3c0766e603624.slice. Oct 28 05:18:13.920984 kubelet[2386]: I1028 05:18:13.920966 2386 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Oct 28 05:18:13.921070 kubelet[2386]: E1028 05:18:13.921015 2386 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Oct 28 05:18:13.921256 kubelet[2386]: E1028 05:18:13.921236 2386 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.103:6443/api/v1/nodes\": dial tcp 10.0.0.103:6443: connect: connection refused" node="localhost" Oct 28 05:18:13.923309 systemd[1]: Created slice kubepods-burstable-poda1d51be1ff02022474f2598f6e43038f.slice - libcontainer container kubepods-burstable-poda1d51be1ff02022474f2598f6e43038f.slice. Oct 28 05:18:13.924823 kubelet[2386]: E1028 05:18:13.924804 2386 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Oct 28 05:18:13.963294 kubelet[2386]: I1028 05:18:13.963256 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2abbad858f7a7505c8a8b4aa1ce70949-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"2abbad858f7a7505c8a8b4aa1ce70949\") " pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:13.963361 kubelet[2386]: I1028 05:18:13.963297 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:13.963361 kubelet[2386]: I1028 05:18:13.963329 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:13.963439 kubelet[2386]: I1028 05:18:13.963356 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:13.963466 kubelet[2386]: I1028 05:18:13.963438 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a1d51be1ff02022474f2598f6e43038f-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a1d51be1ff02022474f2598f6e43038f\") " pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:13.963487 kubelet[2386]: I1028 05:18:13.963463 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2abbad858f7a7505c8a8b4aa1ce70949-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"2abbad858f7a7505c8a8b4aa1ce70949\") " pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:13.963487 kubelet[2386]: I1028 05:18:13.963480 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2abbad858f7a7505c8a8b4aa1ce70949-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"2abbad858f7a7505c8a8b4aa1ce70949\") " pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:13.963531 kubelet[2386]: I1028 05:18:13.963519 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:13.963552 kubelet[2386]: I1028 05:18:13.963543 2386 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:13.988697 kubelet[2386]: W1028 05:18:13.988666 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.103:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:13.988761 kubelet[2386]: E1028 05:18:13.988716 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.103:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:14.216880 kubelet[2386]: E1028 05:18:14.216767 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:14.217745 containerd[1617]: time="2025-10-28T05:18:14.217697630Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:2abbad858f7a7505c8a8b4aa1ce70949,Namespace:kube-system,Attempt:0,}" Oct 28 05:18:14.222323 kubelet[2386]: E1028 05:18:14.222290 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:14.222777 containerd[1617]: time="2025-10-28T05:18:14.222746263Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:4654b122dbb389158fe3c0766e603624,Namespace:kube-system,Attempt:0,}" Oct 28 05:18:14.226022 kubelet[2386]: E1028 05:18:14.225994 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:14.226327 containerd[1617]: time="2025-10-28T05:18:14.226270439Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a1d51be1ff02022474f2598f6e43038f,Namespace:kube-system,Attempt:0,}" Oct 28 05:18:14.323070 kubelet[2386]: I1028 05:18:14.323045 2386 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Oct 28 05:18:14.323552 kubelet[2386]: E1028 05:18:14.323511 2386 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.103:6443/api/v1/nodes\": dial tcp 10.0.0.103:6443: connect: connection refused" node="localhost" Oct 28 05:18:14.423429 kubelet[2386]: W1028 05:18:14.423342 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.103:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:14.423557 kubelet[2386]: E1028 05:18:14.423445 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.103:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:14.563051 kubelet[2386]: E1028 05:18:14.562932 2386 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.103:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.103:6443: connect: connection refused" interval="1.6s" Oct 28 05:18:14.730647 kubelet[2386]: W1028 05:18:14.730585 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.103:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:14.731007 kubelet[2386]: E1028 05:18:14.730649 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.103:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:14.741201 kubelet[2386]: W1028 05:18:14.741164 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.103:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:14.741254 kubelet[2386]: E1028 05:18:14.741200 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.103:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:15.125751 kubelet[2386]: I1028 05:18:15.125707 2386 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Oct 28 05:18:15.126173 kubelet[2386]: E1028 05:18:15.126139 2386 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.103:6443/api/v1/nodes\": dial tcp 10.0.0.103:6443: connect: connection refused" node="localhost" Oct 28 05:18:15.245890 kubelet[2386]: E1028 05:18:15.245848 2386 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.103:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:15.965151 kubelet[2386]: W1028 05:18:15.965092 2386 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.103:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.103:6443: connect: connection refused Oct 28 05:18:15.965151 kubelet[2386]: E1028 05:18:15.965149 2386 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.103:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.103:6443: connect: connection refused" logger="UnhandledError" Oct 28 05:18:16.119290 containerd[1617]: time="2025-10-28T05:18:16.119195587Z" level=info msg="connecting to shim 931ec831f1934355cc49c08f9ad8df0ddff7c5d397ff04fc855c4cc5ffe929d5" address="unix:///run/containerd/s/dbb04055c322e4320185cafd04d74189cedacc35500ffd365895b80759c7c187" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:16.132205 containerd[1617]: time="2025-10-28T05:18:16.132134947Z" level=info msg="connecting to shim cdf341009fd157263686bf22094f9d75df4cc417fd334d21399dc852589e1cab" address="unix:///run/containerd/s/84f9ea6d0d9bff6e196895fbc835af0fd52192dcc9c853c671c4dea7be848398" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:16.154755 containerd[1617]: time="2025-10-28T05:18:16.154656955Z" level=info msg="connecting to shim 6627324e0679888f4747ab28e426941f9dfb4e016fc85f5093e68d680a620b44" address="unix:///run/containerd/s/fa3ee7efc38acfefaecde9e4bac6d89d34e0ddce793b2e7776f5dc8cbc19c115" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:16.163832 kubelet[2386]: E1028 05:18:16.163760 2386 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.103:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.103:6443: connect: connection refused" interval="3.2s" Oct 28 05:18:16.166548 systemd[1]: Started cri-containerd-cdf341009fd157263686bf22094f9d75df4cc417fd334d21399dc852589e1cab.scope - libcontainer container cdf341009fd157263686bf22094f9d75df4cc417fd334d21399dc852589e1cab. Oct 28 05:18:16.170726 systemd[1]: Started cri-containerd-931ec831f1934355cc49c08f9ad8df0ddff7c5d397ff04fc855c4cc5ffe929d5.scope - libcontainer container 931ec831f1934355cc49c08f9ad8df0ddff7c5d397ff04fc855c4cc5ffe929d5. Oct 28 05:18:16.190545 systemd[1]: Started cri-containerd-6627324e0679888f4747ab28e426941f9dfb4e016fc85f5093e68d680a620b44.scope - libcontainer container 6627324e0679888f4747ab28e426941f9dfb4e016fc85f5093e68d680a620b44. Oct 28 05:18:16.258543 containerd[1617]: time="2025-10-28T05:18:16.258269389Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:4654b122dbb389158fe3c0766e603624,Namespace:kube-system,Attempt:0,} returns sandbox id \"931ec831f1934355cc49c08f9ad8df0ddff7c5d397ff04fc855c4cc5ffe929d5\"" Oct 28 05:18:16.270234 kubelet[2386]: E1028 05:18:16.269938 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:16.273044 containerd[1617]: time="2025-10-28T05:18:16.272990500Z" level=info msg="CreateContainer within sandbox \"931ec831f1934355cc49c08f9ad8df0ddff7c5d397ff04fc855c4cc5ffe929d5\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Oct 28 05:18:16.285614 containerd[1617]: time="2025-10-28T05:18:16.285567206Z" level=info msg="Container a268fe113dc82d0aa688459ba43d0d0b1d4fe7dcd70d449b830b2814d84985c3: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:16.295706 containerd[1617]: time="2025-10-28T05:18:16.295663227Z" level=info msg="CreateContainer within sandbox \"931ec831f1934355cc49c08f9ad8df0ddff7c5d397ff04fc855c4cc5ffe929d5\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"a268fe113dc82d0aa688459ba43d0d0b1d4fe7dcd70d449b830b2814d84985c3\"" Oct 28 05:18:16.296891 containerd[1617]: time="2025-10-28T05:18:16.296706569Z" level=info msg="StartContainer for \"a268fe113dc82d0aa688459ba43d0d0b1d4fe7dcd70d449b830b2814d84985c3\"" Oct 28 05:18:16.298079 containerd[1617]: time="2025-10-28T05:18:16.298057808Z" level=info msg="connecting to shim a268fe113dc82d0aa688459ba43d0d0b1d4fe7dcd70d449b830b2814d84985c3" address="unix:///run/containerd/s/dbb04055c322e4320185cafd04d74189cedacc35500ffd365895b80759c7c187" protocol=ttrpc version=3 Oct 28 05:18:16.321380 containerd[1617]: time="2025-10-28T05:18:16.321315441Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:2abbad858f7a7505c8a8b4aa1ce70949,Namespace:kube-system,Attempt:0,} returns sandbox id \"cdf341009fd157263686bf22094f9d75df4cc417fd334d21399dc852589e1cab\"" Oct 28 05:18:16.322121 kubelet[2386]: E1028 05:18:16.322098 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:16.324691 containerd[1617]: time="2025-10-28T05:18:16.323913017Z" level=info msg="CreateContainer within sandbox \"cdf341009fd157263686bf22094f9d75df4cc417fd334d21399dc852589e1cab\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Oct 28 05:18:16.325665 systemd[1]: Started cri-containerd-a268fe113dc82d0aa688459ba43d0d0b1d4fe7dcd70d449b830b2814d84985c3.scope - libcontainer container a268fe113dc82d0aa688459ba43d0d0b1d4fe7dcd70d449b830b2814d84985c3. Oct 28 05:18:16.336778 containerd[1617]: time="2025-10-28T05:18:16.336739020Z" level=info msg="Container db9421d9737beb35689b7c467b1ac810b51f75b086adb11c3d2eba8129c6d2ae: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:16.337681 containerd[1617]: time="2025-10-28T05:18:16.337646445Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a1d51be1ff02022474f2598f6e43038f,Namespace:kube-system,Attempt:0,} returns sandbox id \"6627324e0679888f4747ab28e426941f9dfb4e016fc85f5093e68d680a620b44\"" Oct 28 05:18:16.339212 kubelet[2386]: E1028 05:18:16.339187 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:16.341687 containerd[1617]: time="2025-10-28T05:18:16.341646514Z" level=info msg="CreateContainer within sandbox \"6627324e0679888f4747ab28e426941f9dfb4e016fc85f5093e68d680a620b44\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Oct 28 05:18:16.344196 containerd[1617]: time="2025-10-28T05:18:16.344157325Z" level=info msg="CreateContainer within sandbox \"cdf341009fd157263686bf22094f9d75df4cc417fd334d21399dc852589e1cab\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"db9421d9737beb35689b7c467b1ac810b51f75b086adb11c3d2eba8129c6d2ae\"" Oct 28 05:18:16.344704 containerd[1617]: time="2025-10-28T05:18:16.344677665Z" level=info msg="StartContainer for \"db9421d9737beb35689b7c467b1ac810b51f75b086adb11c3d2eba8129c6d2ae\"" Oct 28 05:18:16.346028 containerd[1617]: time="2025-10-28T05:18:16.345990020Z" level=info msg="connecting to shim db9421d9737beb35689b7c467b1ac810b51f75b086adb11c3d2eba8129c6d2ae" address="unix:///run/containerd/s/84f9ea6d0d9bff6e196895fbc835af0fd52192dcc9c853c671c4dea7be848398" protocol=ttrpc version=3 Oct 28 05:18:16.356213 containerd[1617]: time="2025-10-28T05:18:16.356166081Z" level=info msg="Container 601a0002250e75846d679fcaa66bd3fc7c458837c9d09678feb668dd2558d791: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:16.363798 containerd[1617]: time="2025-10-28T05:18:16.363758667Z" level=info msg="CreateContainer within sandbox \"6627324e0679888f4747ab28e426941f9dfb4e016fc85f5093e68d680a620b44\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"601a0002250e75846d679fcaa66bd3fc7c458837c9d09678feb668dd2558d791\"" Oct 28 05:18:16.364152 containerd[1617]: time="2025-10-28T05:18:16.364117657Z" level=info msg="StartContainer for \"601a0002250e75846d679fcaa66bd3fc7c458837c9d09678feb668dd2558d791\"" Oct 28 05:18:16.365227 containerd[1617]: time="2025-10-28T05:18:16.365200333Z" level=info msg="connecting to shim 601a0002250e75846d679fcaa66bd3fc7c458837c9d09678feb668dd2558d791" address="unix:///run/containerd/s/fa3ee7efc38acfefaecde9e4bac6d89d34e0ddce793b2e7776f5dc8cbc19c115" protocol=ttrpc version=3 Oct 28 05:18:16.378583 systemd[1]: Started cri-containerd-db9421d9737beb35689b7c467b1ac810b51f75b086adb11c3d2eba8129c6d2ae.scope - libcontainer container db9421d9737beb35689b7c467b1ac810b51f75b086adb11c3d2eba8129c6d2ae. Oct 28 05:18:16.457575 systemd[1]: Started cri-containerd-601a0002250e75846d679fcaa66bd3fc7c458837c9d09678feb668dd2558d791.scope - libcontainer container 601a0002250e75846d679fcaa66bd3fc7c458837c9d09678feb668dd2558d791. Oct 28 05:18:16.466319 containerd[1617]: time="2025-10-28T05:18:16.466216492Z" level=info msg="StartContainer for \"a268fe113dc82d0aa688459ba43d0d0b1d4fe7dcd70d449b830b2814d84985c3\" returns successfully" Oct 28 05:18:16.514835 containerd[1617]: time="2025-10-28T05:18:16.514698541Z" level=info msg="StartContainer for \"db9421d9737beb35689b7c467b1ac810b51f75b086adb11c3d2eba8129c6d2ae\" returns successfully" Oct 28 05:18:16.523265 containerd[1617]: time="2025-10-28T05:18:16.523215025Z" level=info msg="StartContainer for \"601a0002250e75846d679fcaa66bd3fc7c458837c9d09678feb668dd2558d791\" returns successfully" Oct 28 05:18:16.730204 kubelet[2386]: I1028 05:18:16.730035 2386 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Oct 28 05:18:17.194745 kubelet[2386]: E1028 05:18:17.194698 2386 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Oct 28 05:18:17.195227 kubelet[2386]: E1028 05:18:17.194836 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:17.197736 kubelet[2386]: E1028 05:18:17.197704 2386 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Oct 28 05:18:17.197855 kubelet[2386]: E1028 05:18:17.197832 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:17.199024 kubelet[2386]: E1028 05:18:17.198992 2386 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Oct 28 05:18:17.199124 kubelet[2386]: E1028 05:18:17.199104 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:18.071251 kubelet[2386]: I1028 05:18:18.071200 2386 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Oct 28 05:18:18.152324 kubelet[2386]: I1028 05:18:18.151373 2386 apiserver.go:52] "Watching apiserver" Oct 28 05:18:18.159806 kubelet[2386]: I1028 05:18:18.159384 2386 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:18.213089 kubelet[2386]: I1028 05:18:18.213050 2386 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:18.213583 kubelet[2386]: I1028 05:18:18.213254 2386 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:18.221144 kubelet[2386]: E1028 05:18:18.221093 2386 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:18.221332 kubelet[2386]: E1028 05:18:18.221288 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:18.221442 kubelet[2386]: E1028 05:18:18.221388 2386 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:18.221482 kubelet[2386]: I1028 05:18:18.221449 2386 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:18.221653 kubelet[2386]: E1028 05:18:18.221630 2386 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:18.221744 kubelet[2386]: E1028 05:18:18.221725 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:18.223604 kubelet[2386]: E1028 05:18:18.223575 2386 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:18.223740 kubelet[2386]: I1028 05:18:18.223677 2386 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:18.225422 kubelet[2386]: E1028 05:18:18.225383 2386 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:18.259796 kubelet[2386]: I1028 05:18:18.259742 2386 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 28 05:18:19.215096 kubelet[2386]: I1028 05:18:19.215050 2386 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:19.219698 kubelet[2386]: E1028 05:18:19.219645 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:19.989277 systemd[1]: Reload requested from client PID 2661 ('systemctl') (unit session-7.scope)... Oct 28 05:18:19.989295 systemd[1]: Reloading... Oct 28 05:18:20.082440 zram_generator::config[2709]: No configuration found. Oct 28 05:18:20.217348 kubelet[2386]: E1028 05:18:20.217252 2386 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:20.320363 systemd[1]: Reloading finished in 330 ms. Oct 28 05:18:20.349520 kubelet[2386]: I1028 05:18:20.349477 2386 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 28 05:18:20.349733 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Oct 28 05:18:20.361719 systemd[1]: kubelet.service: Deactivated successfully. Oct 28 05:18:20.362087 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 28 05:18:20.362143 systemd[1]: kubelet.service: Consumed 1.131s CPU time, 130.9M memory peak. Oct 28 05:18:20.364054 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 28 05:18:20.587810 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 28 05:18:20.596801 (kubelet)[2750]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 28 05:18:20.641085 kubelet[2750]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 28 05:18:20.641085 kubelet[2750]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 28 05:18:20.641085 kubelet[2750]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 28 05:18:20.641530 kubelet[2750]: I1028 05:18:20.641127 2750 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 28 05:18:20.649656 kubelet[2750]: I1028 05:18:20.649598 2750 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 28 05:18:20.649656 kubelet[2750]: I1028 05:18:20.649630 2750 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 28 05:18:20.649958 kubelet[2750]: I1028 05:18:20.649921 2750 server.go:954] "Client rotation is on, will bootstrap in background" Oct 28 05:18:20.651270 kubelet[2750]: I1028 05:18:20.651224 2750 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 28 05:18:20.653471 kubelet[2750]: I1028 05:18:20.653428 2750 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 28 05:18:20.657867 kubelet[2750]: I1028 05:18:20.657840 2750 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 28 05:18:20.664711 kubelet[2750]: I1028 05:18:20.663557 2750 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 28 05:18:20.664711 kubelet[2750]: I1028 05:18:20.663804 2750 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 28 05:18:20.664711 kubelet[2750]: I1028 05:18:20.663826 2750 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 28 05:18:20.664711 kubelet[2750]: I1028 05:18:20.664103 2750 topology_manager.go:138] "Creating topology manager with none policy" Oct 28 05:18:20.664929 kubelet[2750]: I1028 05:18:20.664112 2750 container_manager_linux.go:304] "Creating device plugin manager" Oct 28 05:18:20.664929 kubelet[2750]: I1028 05:18:20.664164 2750 state_mem.go:36] "Initialized new in-memory state store" Oct 28 05:18:20.664929 kubelet[2750]: I1028 05:18:20.664324 2750 kubelet.go:446] "Attempting to sync node with API server" Oct 28 05:18:20.664929 kubelet[2750]: I1028 05:18:20.664347 2750 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 28 05:18:20.664929 kubelet[2750]: I1028 05:18:20.664372 2750 kubelet.go:352] "Adding apiserver pod source" Oct 28 05:18:20.664929 kubelet[2750]: I1028 05:18:20.664382 2750 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 28 05:18:20.665533 kubelet[2750]: I1028 05:18:20.665494 2750 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Oct 28 05:18:20.666154 kubelet[2750]: I1028 05:18:20.666138 2750 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 28 05:18:20.667929 kubelet[2750]: I1028 05:18:20.667894 2750 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 28 05:18:20.667982 kubelet[2750]: I1028 05:18:20.667942 2750 server.go:1287] "Started kubelet" Oct 28 05:18:20.669707 kubelet[2750]: I1028 05:18:20.669665 2750 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 28 05:18:20.670505 kubelet[2750]: I1028 05:18:20.670261 2750 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 28 05:18:20.673269 kubelet[2750]: I1028 05:18:20.672441 2750 server.go:479] "Adding debug handlers to kubelet server" Oct 28 05:18:20.673269 kubelet[2750]: I1028 05:18:20.673024 2750 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 28 05:18:20.673269 kubelet[2750]: E1028 05:18:20.673129 2750 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 28 05:18:20.673363 kubelet[2750]: I1028 05:18:20.673306 2750 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 28 05:18:20.673512 kubelet[2750]: I1028 05:18:20.673478 2750 reconciler.go:26] "Reconciler: start to sync state" Oct 28 05:18:20.674345 kubelet[2750]: I1028 05:18:20.670899 2750 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 28 05:18:20.674345 kubelet[2750]: I1028 05:18:20.670435 2750 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 28 05:18:20.674572 kubelet[2750]: I1028 05:18:20.674545 2750 factory.go:221] Registration of the systemd container factory successfully Oct 28 05:18:20.674672 kubelet[2750]: I1028 05:18:20.674644 2750 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 28 05:18:20.675177 kubelet[2750]: I1028 05:18:20.675149 2750 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 28 05:18:20.676645 kubelet[2750]: I1028 05:18:20.676613 2750 factory.go:221] Registration of the containerd container factory successfully Oct 28 05:18:20.678965 kubelet[2750]: E1028 05:18:20.678932 2750 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 28 05:18:20.691084 kubelet[2750]: I1028 05:18:20.690529 2750 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 28 05:18:20.691865 kubelet[2750]: I1028 05:18:20.691841 2750 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 28 05:18:20.691979 kubelet[2750]: I1028 05:18:20.691968 2750 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 28 05:18:20.692058 kubelet[2750]: I1028 05:18:20.692040 2750 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 28 05:18:20.692110 kubelet[2750]: I1028 05:18:20.692102 2750 kubelet.go:2382] "Starting kubelet main sync loop" Oct 28 05:18:20.692251 kubelet[2750]: E1028 05:18:20.692216 2750 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 28 05:18:20.722618 kubelet[2750]: I1028 05:18:20.722582 2750 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 28 05:18:20.722618 kubelet[2750]: I1028 05:18:20.722601 2750 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 28 05:18:20.722618 kubelet[2750]: I1028 05:18:20.722624 2750 state_mem.go:36] "Initialized new in-memory state store" Oct 28 05:18:20.722809 kubelet[2750]: I1028 05:18:20.722765 2750 state_mem.go:88] "Updated default CPUSet" cpuSet="" Oct 28 05:18:20.722809 kubelet[2750]: I1028 05:18:20.722781 2750 state_mem.go:96] "Updated CPUSet assignments" assignments={} Oct 28 05:18:20.722809 kubelet[2750]: I1028 05:18:20.722799 2750 policy_none.go:49] "None policy: Start" Oct 28 05:18:20.722809 kubelet[2750]: I1028 05:18:20.722808 2750 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 28 05:18:20.722904 kubelet[2750]: I1028 05:18:20.722819 2750 state_mem.go:35] "Initializing new in-memory state store" Oct 28 05:18:20.722997 kubelet[2750]: I1028 05:18:20.722929 2750 state_mem.go:75] "Updated machine memory state" Oct 28 05:18:20.727263 kubelet[2750]: I1028 05:18:20.727090 2750 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 28 05:18:20.727328 kubelet[2750]: I1028 05:18:20.727267 2750 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 28 05:18:20.727328 kubelet[2750]: I1028 05:18:20.727280 2750 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 28 05:18:20.727520 kubelet[2750]: I1028 05:18:20.727497 2750 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 28 05:18:20.728690 kubelet[2750]: E1028 05:18:20.728667 2750 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 28 05:18:20.793659 kubelet[2750]: I1028 05:18:20.793592 2750 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:20.793778 kubelet[2750]: I1028 05:18:20.793684 2750 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:20.793807 kubelet[2750]: I1028 05:18:20.793779 2750 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:20.799853 kubelet[2750]: E1028 05:18:20.799792 2750 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:20.833408 kubelet[2750]: I1028 05:18:20.833349 2750 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Oct 28 05:18:20.839212 kubelet[2750]: I1028 05:18:20.839122 2750 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Oct 28 05:18:20.839212 kubelet[2750]: I1028 05:18:20.839182 2750 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Oct 28 05:18:20.975221 kubelet[2750]: I1028 05:18:20.975176 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2abbad858f7a7505c8a8b4aa1ce70949-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"2abbad858f7a7505c8a8b4aa1ce70949\") " pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:20.975221 kubelet[2750]: I1028 05:18:20.975214 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2abbad858f7a7505c8a8b4aa1ce70949-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"2abbad858f7a7505c8a8b4aa1ce70949\") " pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:20.975221 kubelet[2750]: I1028 05:18:20.975236 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:20.975476 kubelet[2750]: I1028 05:18:20.975268 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:20.975476 kubelet[2750]: I1028 05:18:20.975315 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:20.975476 kubelet[2750]: I1028 05:18:20.975337 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:20.975476 kubelet[2750]: I1028 05:18:20.975358 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2abbad858f7a7505c8a8b4aa1ce70949-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"2abbad858f7a7505c8a8b4aa1ce70949\") " pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:20.975476 kubelet[2750]: I1028 05:18:20.975382 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Oct 28 05:18:20.975651 kubelet[2750]: I1028 05:18:20.975440 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a1d51be1ff02022474f2598f6e43038f-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a1d51be1ff02022474f2598f6e43038f\") " pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:21.101256 kubelet[2750]: E1028 05:18:21.100985 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:21.101256 kubelet[2750]: E1028 05:18:21.101119 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:21.101256 kubelet[2750]: E1028 05:18:21.101129 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:21.666560 kubelet[2750]: I1028 05:18:21.666489 2750 apiserver.go:52] "Watching apiserver" Oct 28 05:18:21.673907 kubelet[2750]: I1028 05:18:21.673861 2750 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 28 05:18:21.709304 kubelet[2750]: I1028 05:18:21.705821 2750 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:21.709304 kubelet[2750]: I1028 05:18:21.707219 2750 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:21.709612 kubelet[2750]: E1028 05:18:21.709549 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:21.712288 kubelet[2750]: E1028 05:18:21.712261 2750 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Oct 28 05:18:21.712540 kubelet[2750]: E1028 05:18:21.712510 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:21.713604 kubelet[2750]: E1028 05:18:21.713583 2750 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Oct 28 05:18:21.713770 kubelet[2750]: E1028 05:18:21.713754 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:21.798181 kubelet[2750]: I1028 05:18:21.797730 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.797709891 podStartE2EDuration="1.797709891s" podCreationTimestamp="2025-10-28 05:18:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-28 05:18:21.732642041 +0000 UTC m=+1.131625012" watchObservedRunningTime="2025-10-28 05:18:21.797709891 +0000 UTC m=+1.196692862" Oct 28 05:18:21.819426 kubelet[2750]: I1028 05:18:21.819162 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.819141773 podStartE2EDuration="2.819141773s" podCreationTimestamp="2025-10-28 05:18:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-28 05:18:21.798054255 +0000 UTC m=+1.197037226" watchObservedRunningTime="2025-10-28 05:18:21.819141773 +0000 UTC m=+1.218124744" Oct 28 05:18:21.819426 kubelet[2750]: I1028 05:18:21.819308 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.819302156 podStartE2EDuration="1.819302156s" podCreationTimestamp="2025-10-28 05:18:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-28 05:18:21.818971808 +0000 UTC m=+1.217954789" watchObservedRunningTime="2025-10-28 05:18:21.819302156 +0000 UTC m=+1.218285127" Oct 28 05:18:22.707746 kubelet[2750]: E1028 05:18:22.707690 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:22.708203 kubelet[2750]: E1028 05:18:22.707862 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:23.709327 kubelet[2750]: E1028 05:18:23.709287 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:25.781618 kubelet[2750]: E1028 05:18:25.781577 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:26.629419 kubelet[2750]: E1028 05:18:26.629367 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:26.713158 kubelet[2750]: E1028 05:18:26.713126 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:26.990913 kubelet[2750]: I1028 05:18:26.990733 2750 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Oct 28 05:18:26.991609 kubelet[2750]: I1028 05:18:26.991361 2750 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Oct 28 05:18:26.991646 containerd[1617]: time="2025-10-28T05:18:26.991128655Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Oct 28 05:18:27.714686 kubelet[2750]: E1028 05:18:27.714635 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:27.774527 systemd[1]: Created slice kubepods-besteffort-pod71c8974f_4279_4d96_9b83_da1afecc5ab3.slice - libcontainer container kubepods-besteffort-pod71c8974f_4279_4d96_9b83_da1afecc5ab3.slice. Oct 28 05:18:27.782767 kubelet[2750]: E1028 05:18:27.782721 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:27.819344 kubelet[2750]: I1028 05:18:27.819138 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/71c8974f-4279-4d96-9b83-da1afecc5ab3-kube-proxy\") pod \"kube-proxy-gszkg\" (UID: \"71c8974f-4279-4d96-9b83-da1afecc5ab3\") " pod="kube-system/kube-proxy-gszkg" Oct 28 05:18:27.819344 kubelet[2750]: I1028 05:18:27.819198 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/71c8974f-4279-4d96-9b83-da1afecc5ab3-xtables-lock\") pod \"kube-proxy-gszkg\" (UID: \"71c8974f-4279-4d96-9b83-da1afecc5ab3\") " pod="kube-system/kube-proxy-gszkg" Oct 28 05:18:27.819344 kubelet[2750]: I1028 05:18:27.819231 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71c8974f-4279-4d96-9b83-da1afecc5ab3-lib-modules\") pod \"kube-proxy-gszkg\" (UID: \"71c8974f-4279-4d96-9b83-da1afecc5ab3\") " pod="kube-system/kube-proxy-gszkg" Oct 28 05:18:27.819344 kubelet[2750]: I1028 05:18:27.819249 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp7gp\" (UniqueName: \"kubernetes.io/projected/71c8974f-4279-4d96-9b83-da1afecc5ab3-kube-api-access-gp7gp\") pod \"kube-proxy-gszkg\" (UID: \"71c8974f-4279-4d96-9b83-da1afecc5ab3\") " pod="kube-system/kube-proxy-gszkg" Oct 28 05:18:27.925256 kubelet[2750]: E1028 05:18:27.925197 2750 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Oct 28 05:18:27.925256 kubelet[2750]: E1028 05:18:27.925241 2750 projected.go:194] Error preparing data for projected volume kube-api-access-gp7gp for pod kube-system/kube-proxy-gszkg: configmap "kube-root-ca.crt" not found Oct 28 05:18:27.925474 kubelet[2750]: E1028 05:18:27.925299 2750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71c8974f-4279-4d96-9b83-da1afecc5ab3-kube-api-access-gp7gp podName:71c8974f-4279-4d96-9b83-da1afecc5ab3 nodeName:}" failed. No retries permitted until 2025-10-28 05:18:28.425279468 +0000 UTC m=+7.824262439 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gp7gp" (UniqueName: "kubernetes.io/projected/71c8974f-4279-4d96-9b83-da1afecc5ab3-kube-api-access-gp7gp") pod "kube-proxy-gszkg" (UID: "71c8974f-4279-4d96-9b83-da1afecc5ab3") : configmap "kube-root-ca.crt" not found Oct 28 05:18:28.085677 systemd[1]: Created slice kubepods-besteffort-podc4222183_0eef_48ca_af33_b17e76107122.slice - libcontainer container kubepods-besteffort-podc4222183_0eef_48ca_af33_b17e76107122.slice. Oct 28 05:18:28.121575 kubelet[2750]: I1028 05:18:28.121386 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/c4222183-0eef-48ca-af33-b17e76107122-var-lib-calico\") pod \"tigera-operator-7dcd859c48-n24j2\" (UID: \"c4222183-0eef-48ca-af33-b17e76107122\") " pod="tigera-operator/tigera-operator-7dcd859c48-n24j2" Oct 28 05:18:28.121575 kubelet[2750]: I1028 05:18:28.121476 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75kzx\" (UniqueName: \"kubernetes.io/projected/c4222183-0eef-48ca-af33-b17e76107122-kube-api-access-75kzx\") pod \"tigera-operator-7dcd859c48-n24j2\" (UID: \"c4222183-0eef-48ca-af33-b17e76107122\") " pod="tigera-operator/tigera-operator-7dcd859c48-n24j2" Oct 28 05:18:28.391539 containerd[1617]: time="2025-10-28T05:18:28.391354955Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-n24j2,Uid:c4222183-0eef-48ca-af33-b17e76107122,Namespace:tigera-operator,Attempt:0,}" Oct 28 05:18:28.414434 containerd[1617]: time="2025-10-28T05:18:28.414190287Z" level=info msg="connecting to shim 8855e2fb7f68d34e0276a7d2cc8e461d77d504a67464c29e99c6d05320a80d4d" address="unix:///run/containerd/s/525fe43b0018303c1ff9be6b18559a1ed905fcc0e70475cd6fad4d263f5632fc" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:28.448566 systemd[1]: Started cri-containerd-8855e2fb7f68d34e0276a7d2cc8e461d77d504a67464c29e99c6d05320a80d4d.scope - libcontainer container 8855e2fb7f68d34e0276a7d2cc8e461d77d504a67464c29e99c6d05320a80d4d. Oct 28 05:18:28.494933 containerd[1617]: time="2025-10-28T05:18:28.494846766Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-n24j2,Uid:c4222183-0eef-48ca-af33-b17e76107122,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"8855e2fb7f68d34e0276a7d2cc8e461d77d504a67464c29e99c6d05320a80d4d\"" Oct 28 05:18:28.496520 containerd[1617]: time="2025-10-28T05:18:28.496494340Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Oct 28 05:18:28.685567 kubelet[2750]: E1028 05:18:28.685434 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:28.685916 containerd[1617]: time="2025-10-28T05:18:28.685833659Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gszkg,Uid:71c8974f-4279-4d96-9b83-da1afecc5ab3,Namespace:kube-system,Attempt:0,}" Oct 28 05:18:28.709531 containerd[1617]: time="2025-10-28T05:18:28.709484533Z" level=info msg="connecting to shim 6f76cb285bcff6b6f173088e3fcc2fe2cbd76b1fe7076ae9fd78e6ed1e041ff3" address="unix:///run/containerd/s/6173dde7ab0429216ca6fa49b10203809493514c5c04f2db393148b0c14bd3c2" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:28.718037 kubelet[2750]: E1028 05:18:28.717762 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:28.749601 systemd[1]: Started cri-containerd-6f76cb285bcff6b6f173088e3fcc2fe2cbd76b1fe7076ae9fd78e6ed1e041ff3.scope - libcontainer container 6f76cb285bcff6b6f173088e3fcc2fe2cbd76b1fe7076ae9fd78e6ed1e041ff3. Oct 28 05:18:28.779984 containerd[1617]: time="2025-10-28T05:18:28.779941643Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gszkg,Uid:71c8974f-4279-4d96-9b83-da1afecc5ab3,Namespace:kube-system,Attempt:0,} returns sandbox id \"6f76cb285bcff6b6f173088e3fcc2fe2cbd76b1fe7076ae9fd78e6ed1e041ff3\"" Oct 28 05:18:28.780735 kubelet[2750]: E1028 05:18:28.780711 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:28.783363 containerd[1617]: time="2025-10-28T05:18:28.783025379Z" level=info msg="CreateContainer within sandbox \"6f76cb285bcff6b6f173088e3fcc2fe2cbd76b1fe7076ae9fd78e6ed1e041ff3\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Oct 28 05:18:28.794423 containerd[1617]: time="2025-10-28T05:18:28.794377403Z" level=info msg="Container 131243fb490a05b98d6ed84b5fdecaa7311c92d50a9ccb126543025acbe36f81: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:28.802924 containerd[1617]: time="2025-10-28T05:18:28.802876415Z" level=info msg="CreateContainer within sandbox \"6f76cb285bcff6b6f173088e3fcc2fe2cbd76b1fe7076ae9fd78e6ed1e041ff3\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"131243fb490a05b98d6ed84b5fdecaa7311c92d50a9ccb126543025acbe36f81\"" Oct 28 05:18:28.804389 containerd[1617]: time="2025-10-28T05:18:28.803389269Z" level=info msg="StartContainer for \"131243fb490a05b98d6ed84b5fdecaa7311c92d50a9ccb126543025acbe36f81\"" Oct 28 05:18:28.804805 containerd[1617]: time="2025-10-28T05:18:28.804760612Z" level=info msg="connecting to shim 131243fb490a05b98d6ed84b5fdecaa7311c92d50a9ccb126543025acbe36f81" address="unix:///run/containerd/s/6173dde7ab0429216ca6fa49b10203809493514c5c04f2db393148b0c14bd3c2" protocol=ttrpc version=3 Oct 28 05:18:28.828543 systemd[1]: Started cri-containerd-131243fb490a05b98d6ed84b5fdecaa7311c92d50a9ccb126543025acbe36f81.scope - libcontainer container 131243fb490a05b98d6ed84b5fdecaa7311c92d50a9ccb126543025acbe36f81. Oct 28 05:18:28.878368 containerd[1617]: time="2025-10-28T05:18:28.878305765Z" level=info msg="StartContainer for \"131243fb490a05b98d6ed84b5fdecaa7311c92d50a9ccb126543025acbe36f81\" returns successfully" Oct 28 05:18:29.721408 kubelet[2750]: E1028 05:18:29.721346 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:29.729731 kubelet[2750]: I1028 05:18:29.729670 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-gszkg" podStartSLOduration=2.729638871 podStartE2EDuration="2.729638871s" podCreationTimestamp="2025-10-28 05:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-28 05:18:29.72923547 +0000 UTC m=+9.128218441" watchObservedRunningTime="2025-10-28 05:18:29.729638871 +0000 UTC m=+9.128621832" Oct 28 05:18:30.150314 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4057058521.mount: Deactivated successfully. Oct 28 05:18:30.480077 containerd[1617]: time="2025-10-28T05:18:30.479926981Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:30.480648 containerd[1617]: time="2025-10-28T05:18:30.480617703Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=25061691" Oct 28 05:18:30.481718 containerd[1617]: time="2025-10-28T05:18:30.481687693Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:30.483730 containerd[1617]: time="2025-10-28T05:18:30.483685121Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:30.484269 containerd[1617]: time="2025-10-28T05:18:30.484226545Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 1.987699819s" Oct 28 05:18:30.484326 containerd[1617]: time="2025-10-28T05:18:30.484268797Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Oct 28 05:18:30.486898 containerd[1617]: time="2025-10-28T05:18:30.486862700Z" level=info msg="CreateContainer within sandbox \"8855e2fb7f68d34e0276a7d2cc8e461d77d504a67464c29e99c6d05320a80d4d\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Oct 28 05:18:30.493469 containerd[1617]: time="2025-10-28T05:18:30.493436178Z" level=info msg="Container f5ed6c29c2f9a0a6dd6b5123ac687d910e28d0298e3aac5ed563c62893cff262: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:30.496850 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1038538921.mount: Deactivated successfully. Oct 28 05:18:30.499982 containerd[1617]: time="2025-10-28T05:18:30.499951872Z" level=info msg="CreateContainer within sandbox \"8855e2fb7f68d34e0276a7d2cc8e461d77d504a67464c29e99c6d05320a80d4d\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"f5ed6c29c2f9a0a6dd6b5123ac687d910e28d0298e3aac5ed563c62893cff262\"" Oct 28 05:18:30.500389 containerd[1617]: time="2025-10-28T05:18:30.500354492Z" level=info msg="StartContainer for \"f5ed6c29c2f9a0a6dd6b5123ac687d910e28d0298e3aac5ed563c62893cff262\"" Oct 28 05:18:30.501295 containerd[1617]: time="2025-10-28T05:18:30.501248883Z" level=info msg="connecting to shim f5ed6c29c2f9a0a6dd6b5123ac687d910e28d0298e3aac5ed563c62893cff262" address="unix:///run/containerd/s/525fe43b0018303c1ff9be6b18559a1ed905fcc0e70475cd6fad4d263f5632fc" protocol=ttrpc version=3 Oct 28 05:18:30.562530 systemd[1]: Started cri-containerd-f5ed6c29c2f9a0a6dd6b5123ac687d910e28d0298e3aac5ed563c62893cff262.scope - libcontainer container f5ed6c29c2f9a0a6dd6b5123ac687d910e28d0298e3aac5ed563c62893cff262. Oct 28 05:18:30.594370 containerd[1617]: time="2025-10-28T05:18:30.594332772Z" level=info msg="StartContainer for \"f5ed6c29c2f9a0a6dd6b5123ac687d910e28d0298e3aac5ed563c62893cff262\" returns successfully" Oct 28 05:18:30.733341 kubelet[2750]: I1028 05:18:30.733189 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-n24j2" podStartSLOduration=0.744264707 podStartE2EDuration="2.733169324s" podCreationTimestamp="2025-10-28 05:18:28 +0000 UTC" firstStartedPulling="2025-10-28 05:18:28.496185404 +0000 UTC m=+7.895168365" lastFinishedPulling="2025-10-28 05:18:30.485090011 +0000 UTC m=+9.884072982" observedRunningTime="2025-10-28 05:18:30.733097548 +0000 UTC m=+10.132080519" watchObservedRunningTime="2025-10-28 05:18:30.733169324 +0000 UTC m=+10.132152295" Oct 28 05:18:35.787183 kubelet[2750]: E1028 05:18:35.787053 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:36.074002 update_engine[1598]: I20251028 05:18:36.073884 1598 update_attempter.cc:509] Updating boot flags... Oct 28 05:18:36.227376 sudo[1823]: pam_unix(sudo:session): session closed for user root Oct 28 05:18:36.231433 sshd[1822]: Connection closed by 10.0.0.1 port 36806 Oct 28 05:18:36.232449 sshd-session[1819]: pam_unix(sshd:session): session closed for user core Oct 28 05:18:36.262283 systemd[1]: sshd@6-10.0.0.103:22-10.0.0.1:36806.service: Deactivated successfully. Oct 28 05:18:36.265674 systemd[1]: session-7.scope: Deactivated successfully. Oct 28 05:18:36.266141 systemd[1]: session-7.scope: Consumed 4.826s CPU time, 223M memory peak. Oct 28 05:18:36.269312 systemd-logind[1594]: Session 7 logged out. Waiting for processes to exit. Oct 28 05:18:36.272283 systemd-logind[1594]: Removed session 7. Oct 28 05:18:36.739154 kubelet[2750]: E1028 05:18:36.739101 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:40.328134 systemd[1]: Created slice kubepods-besteffort-pod72531341_5198_4186_85af_7cd51f9b5d10.slice - libcontainer container kubepods-besteffort-pod72531341_5198_4186_85af_7cd51f9b5d10.slice. Oct 28 05:18:40.404675 kubelet[2750]: I1028 05:18:40.404606 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72531341-5198-4186-85af-7cd51f9b5d10-tigera-ca-bundle\") pod \"calico-typha-64bb95c58d-v8r42\" (UID: \"72531341-5198-4186-85af-7cd51f9b5d10\") " pod="calico-system/calico-typha-64bb95c58d-v8r42" Oct 28 05:18:40.404675 kubelet[2750]: I1028 05:18:40.404663 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/72531341-5198-4186-85af-7cd51f9b5d10-typha-certs\") pod \"calico-typha-64bb95c58d-v8r42\" (UID: \"72531341-5198-4186-85af-7cd51f9b5d10\") " pod="calico-system/calico-typha-64bb95c58d-v8r42" Oct 28 05:18:40.404675 kubelet[2750]: I1028 05:18:40.404686 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5tcj\" (UniqueName: \"kubernetes.io/projected/72531341-5198-4186-85af-7cd51f9b5d10-kube-api-access-p5tcj\") pod \"calico-typha-64bb95c58d-v8r42\" (UID: \"72531341-5198-4186-85af-7cd51f9b5d10\") " pod="calico-system/calico-typha-64bb95c58d-v8r42" Oct 28 05:18:40.509966 systemd[1]: Created slice kubepods-besteffort-pod51af4ad7_d74e_45f7_8c48_7dfea57da407.slice - libcontainer container kubepods-besteffort-pod51af4ad7_d74e_45f7_8c48_7dfea57da407.slice. Oct 28 05:18:40.606039 kubelet[2750]: I1028 05:18:40.605860 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-cni-log-dir\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606039 kubelet[2750]: I1028 05:18:40.605941 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-lib-modules\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606039 kubelet[2750]: I1028 05:18:40.605967 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/51af4ad7-d74e-45f7-8c48-7dfea57da407-node-certs\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606243 kubelet[2750]: I1028 05:18:40.606031 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-cni-bin-dir\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606243 kubelet[2750]: I1028 05:18:40.606080 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-var-lib-calico\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606243 kubelet[2750]: I1028 05:18:40.606159 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51af4ad7-d74e-45f7-8c48-7dfea57da407-tigera-ca-bundle\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606243 kubelet[2750]: I1028 05:18:40.606186 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-flexvol-driver-host\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606243 kubelet[2750]: I1028 05:18:40.606236 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-xtables-lock\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606448 kubelet[2750]: I1028 05:18:40.606260 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-policysync\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606448 kubelet[2750]: I1028 05:18:40.606288 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-var-run-calico\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606448 kubelet[2750]: I1028 05:18:40.606310 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/51af4ad7-d74e-45f7-8c48-7dfea57da407-cni-net-dir\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.606448 kubelet[2750]: I1028 05:18:40.606328 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4cxg\" (UniqueName: \"kubernetes.io/projected/51af4ad7-d74e-45f7-8c48-7dfea57da407-kube-api-access-b4cxg\") pod \"calico-node-9k4vr\" (UID: \"51af4ad7-d74e-45f7-8c48-7dfea57da407\") " pod="calico-system/calico-node-9k4vr" Oct 28 05:18:40.631090 kubelet[2750]: E1028 05:18:40.631035 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:40.631857 containerd[1617]: time="2025-10-28T05:18:40.631806274Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-64bb95c58d-v8r42,Uid:72531341-5198-4186-85af-7cd51f9b5d10,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:40.719693 kubelet[2750]: E1028 05:18:40.719603 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.719693 kubelet[2750]: W1028 05:18:40.719638 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.724759 kubelet[2750]: E1028 05:18:40.719721 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.724759 kubelet[2750]: E1028 05:18:40.722862 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.724759 kubelet[2750]: W1028 05:18:40.722886 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.724759 kubelet[2750]: E1028 05:18:40.722910 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.724759 kubelet[2750]: E1028 05:18:40.723904 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.724759 kubelet[2750]: W1028 05:18:40.723919 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.724759 kubelet[2750]: E1028 05:18:40.723931 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.724759 kubelet[2750]: E1028 05:18:40.724096 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:18:40.758160 containerd[1617]: time="2025-10-28T05:18:40.758093031Z" level=info msg="connecting to shim 95ad9e3a99699b6a2bd9e6ec6e6372ae59281e43eb2943b5652e0fa86b7addd7" address="unix:///run/containerd/s/af225b9610014eb3a33c81ce2479aca51df5422aec6f195c2594f3502b1b379d" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:40.789741 systemd[1]: Started cri-containerd-95ad9e3a99699b6a2bd9e6ec6e6372ae59281e43eb2943b5652e0fa86b7addd7.scope - libcontainer container 95ad9e3a99699b6a2bd9e6ec6e6372ae59281e43eb2943b5652e0fa86b7addd7. Oct 28 05:18:40.794414 kubelet[2750]: E1028 05:18:40.794361 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.794414 kubelet[2750]: W1028 05:18:40.794386 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.794553 kubelet[2750]: E1028 05:18:40.794426 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.794691 kubelet[2750]: E1028 05:18:40.794663 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.794691 kubelet[2750]: W1028 05:18:40.794682 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.794758 kubelet[2750]: E1028 05:18:40.794693 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.794973 kubelet[2750]: E1028 05:18:40.794950 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.794973 kubelet[2750]: W1028 05:18:40.794964 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.794973 kubelet[2750]: E1028 05:18:40.794973 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.795222 kubelet[2750]: E1028 05:18:40.795200 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.795222 kubelet[2750]: W1028 05:18:40.795214 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.795281 kubelet[2750]: E1028 05:18:40.795223 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.795442 kubelet[2750]: E1028 05:18:40.795426 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.795442 kubelet[2750]: W1028 05:18:40.795438 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.795500 kubelet[2750]: E1028 05:18:40.795448 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.795655 kubelet[2750]: E1028 05:18:40.795624 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.795655 kubelet[2750]: W1028 05:18:40.795645 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.795655 kubelet[2750]: E1028 05:18:40.795655 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.796290 kubelet[2750]: E1028 05:18:40.796266 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.796290 kubelet[2750]: W1028 05:18:40.796281 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.796290 kubelet[2750]: E1028 05:18:40.796290 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.796521 kubelet[2750]: E1028 05:18:40.796499 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.796521 kubelet[2750]: W1028 05:18:40.796512 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.796521 kubelet[2750]: E1028 05:18:40.796522 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.797557 kubelet[2750]: E1028 05:18:40.797523 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.797597 kubelet[2750]: W1028 05:18:40.797552 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.797597 kubelet[2750]: E1028 05:18:40.797584 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.797861 kubelet[2750]: E1028 05:18:40.797841 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.797861 kubelet[2750]: W1028 05:18:40.797855 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.797907 kubelet[2750]: E1028 05:18:40.797865 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.798057 kubelet[2750]: E1028 05:18:40.798043 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.798092 kubelet[2750]: W1028 05:18:40.798057 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.798092 kubelet[2750]: E1028 05:18:40.798068 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.798298 kubelet[2750]: E1028 05:18:40.798285 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.798320 kubelet[2750]: W1028 05:18:40.798297 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.798320 kubelet[2750]: E1028 05:18:40.798306 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.798554 kubelet[2750]: E1028 05:18:40.798540 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.798578 kubelet[2750]: W1028 05:18:40.798554 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.798578 kubelet[2750]: E1028 05:18:40.798564 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.798786 kubelet[2750]: E1028 05:18:40.798772 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.798786 kubelet[2750]: W1028 05:18:40.798784 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.798830 kubelet[2750]: E1028 05:18:40.798794 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.799142 kubelet[2750]: E1028 05:18:40.799120 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.799142 kubelet[2750]: W1028 05:18:40.799133 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.799185 kubelet[2750]: E1028 05:18:40.799143 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.799359 kubelet[2750]: E1028 05:18:40.799340 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.799359 kubelet[2750]: W1028 05:18:40.799351 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.799435 kubelet[2750]: E1028 05:18:40.799362 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.799708 kubelet[2750]: E1028 05:18:40.799687 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.799708 kubelet[2750]: W1028 05:18:40.799700 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.799756 kubelet[2750]: E1028 05:18:40.799711 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.799923 kubelet[2750]: E1028 05:18:40.799909 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.799923 kubelet[2750]: W1028 05:18:40.799920 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.799979 kubelet[2750]: E1028 05:18:40.799946 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.800207 kubelet[2750]: E1028 05:18:40.800186 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.800207 kubelet[2750]: W1028 05:18:40.800199 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.800256 kubelet[2750]: E1028 05:18:40.800210 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.800417 kubelet[2750]: E1028 05:18:40.800389 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.800417 kubelet[2750]: W1028 05:18:40.800413 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.800465 kubelet[2750]: E1028 05:18:40.800422 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.808888 kubelet[2750]: E1028 05:18:40.808849 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.808888 kubelet[2750]: W1028 05:18:40.808868 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.808888 kubelet[2750]: E1028 05:18:40.808883 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.809072 kubelet[2750]: I1028 05:18:40.808911 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/144455de-c836-43b8-85ee-1c567bc354f0-registration-dir\") pod \"csi-node-driver-tfkvb\" (UID: \"144455de-c836-43b8-85ee-1c567bc354f0\") " pod="calico-system/csi-node-driver-tfkvb" Oct 28 05:18:40.809210 kubelet[2750]: E1028 05:18:40.809177 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.809210 kubelet[2750]: W1028 05:18:40.809195 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.809280 kubelet[2750]: E1028 05:18:40.809218 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.809280 kubelet[2750]: I1028 05:18:40.809233 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/144455de-c836-43b8-85ee-1c567bc354f0-socket-dir\") pod \"csi-node-driver-tfkvb\" (UID: \"144455de-c836-43b8-85ee-1c567bc354f0\") " pod="calico-system/csi-node-driver-tfkvb" Oct 28 05:18:40.809597 kubelet[2750]: E1028 05:18:40.809571 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.809597 kubelet[2750]: W1028 05:18:40.809587 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.809681 kubelet[2750]: E1028 05:18:40.809609 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.809866 kubelet[2750]: E1028 05:18:40.809841 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.809866 kubelet[2750]: W1028 05:18:40.809854 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.809916 kubelet[2750]: E1028 05:18:40.809875 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.810111 kubelet[2750]: E1028 05:18:40.810087 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.810111 kubelet[2750]: W1028 05:18:40.810099 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.810160 kubelet[2750]: E1028 05:18:40.810120 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.810334 kubelet[2750]: E1028 05:18:40.810319 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.810334 kubelet[2750]: W1028 05:18:40.810330 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.810380 kubelet[2750]: E1028 05:18:40.810346 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.810559 kubelet[2750]: E1028 05:18:40.810544 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.810559 kubelet[2750]: W1028 05:18:40.810555 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.810620 kubelet[2750]: E1028 05:18:40.810563 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.810620 kubelet[2750]: I1028 05:18:40.810590 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chrf7\" (UniqueName: \"kubernetes.io/projected/144455de-c836-43b8-85ee-1c567bc354f0-kube-api-access-chrf7\") pod \"csi-node-driver-tfkvb\" (UID: \"144455de-c836-43b8-85ee-1c567bc354f0\") " pod="calico-system/csi-node-driver-tfkvb" Oct 28 05:18:40.810856 kubelet[2750]: E1028 05:18:40.810838 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.810856 kubelet[2750]: W1028 05:18:40.810851 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.810902 kubelet[2750]: E1028 05:18:40.810871 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.810902 kubelet[2750]: I1028 05:18:40.810885 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/144455de-c836-43b8-85ee-1c567bc354f0-kubelet-dir\") pod \"csi-node-driver-tfkvb\" (UID: \"144455de-c836-43b8-85ee-1c567bc354f0\") " pod="calico-system/csi-node-driver-tfkvb" Oct 28 05:18:40.811184 kubelet[2750]: E1028 05:18:40.811155 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.811184 kubelet[2750]: W1028 05:18:40.811173 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.811228 kubelet[2750]: E1028 05:18:40.811203 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.811228 kubelet[2750]: I1028 05:18:40.811224 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/144455de-c836-43b8-85ee-1c567bc354f0-varrun\") pod \"csi-node-driver-tfkvb\" (UID: \"144455de-c836-43b8-85ee-1c567bc354f0\") " pod="calico-system/csi-node-driver-tfkvb" Oct 28 05:18:40.811504 kubelet[2750]: E1028 05:18:40.811486 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.811504 kubelet[2750]: W1028 05:18:40.811499 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.811563 kubelet[2750]: E1028 05:18:40.811512 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.811740 kubelet[2750]: E1028 05:18:40.811716 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.811740 kubelet[2750]: W1028 05:18:40.811729 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.811783 kubelet[2750]: E1028 05:18:40.811749 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.811966 kubelet[2750]: E1028 05:18:40.811949 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.811966 kubelet[2750]: W1028 05:18:40.811961 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.812136 kubelet[2750]: E1028 05:18:40.811981 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.812300 kubelet[2750]: E1028 05:18:40.812284 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.812300 kubelet[2750]: W1028 05:18:40.812296 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.812340 kubelet[2750]: E1028 05:18:40.812316 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.812538 kubelet[2750]: E1028 05:18:40.812524 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.812538 kubelet[2750]: W1028 05:18:40.812535 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.812596 kubelet[2750]: E1028 05:18:40.812544 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.812835 kubelet[2750]: E1028 05:18:40.812810 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.812835 kubelet[2750]: W1028 05:18:40.812825 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.812882 kubelet[2750]: E1028 05:18:40.812835 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.820096 kubelet[2750]: E1028 05:18:40.820073 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:40.820678 containerd[1617]: time="2025-10-28T05:18:40.820622211Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-9k4vr,Uid:51af4ad7-d74e-45f7-8c48-7dfea57da407,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:40.850034 containerd[1617]: time="2025-10-28T05:18:40.849983985Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-64bb95c58d-v8r42,Uid:72531341-5198-4186-85af-7cd51f9b5d10,Namespace:calico-system,Attempt:0,} returns sandbox id \"95ad9e3a99699b6a2bd9e6ec6e6372ae59281e43eb2943b5652e0fa86b7addd7\"" Oct 28 05:18:40.852038 kubelet[2750]: E1028 05:18:40.851961 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:40.854702 containerd[1617]: time="2025-10-28T05:18:40.854659209Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Oct 28 05:18:40.864788 containerd[1617]: time="2025-10-28T05:18:40.864656243Z" level=info msg="connecting to shim d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2" address="unix:///run/containerd/s/ff54ef066755ecfaefbc3f0f98152dc00bfd2bc80aab441ce2153e9cb809ef44" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:40.895555 systemd[1]: Started cri-containerd-d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2.scope - libcontainer container d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2. Oct 28 05:18:40.911866 kubelet[2750]: E1028 05:18:40.911831 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.911866 kubelet[2750]: W1028 05:18:40.911852 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.911866 kubelet[2750]: E1028 05:18:40.911871 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.912130 kubelet[2750]: E1028 05:18:40.912108 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.912130 kubelet[2750]: W1028 05:18:40.912120 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.912186 kubelet[2750]: E1028 05:18:40.912136 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.912348 kubelet[2750]: E1028 05:18:40.912327 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.912348 kubelet[2750]: W1028 05:18:40.912338 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.912390 kubelet[2750]: E1028 05:18:40.912358 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.912629 kubelet[2750]: E1028 05:18:40.912603 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.912629 kubelet[2750]: W1028 05:18:40.912623 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.912693 kubelet[2750]: E1028 05:18:40.912671 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.913203 kubelet[2750]: E1028 05:18:40.913178 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.913309 kubelet[2750]: W1028 05:18:40.913236 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.913378 kubelet[2750]: E1028 05:18:40.913363 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.914182 kubelet[2750]: E1028 05:18:40.914162 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.914276 kubelet[2750]: W1028 05:18:40.914264 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.914384 kubelet[2750]: E1028 05:18:40.914347 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.915628 kubelet[2750]: E1028 05:18:40.915494 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.915628 kubelet[2750]: W1028 05:18:40.915526 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.915628 kubelet[2750]: E1028 05:18:40.915537 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.915789 kubelet[2750]: E1028 05:18:40.915765 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.915921 kubelet[2750]: W1028 05:18:40.915867 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.915983 kubelet[2750]: E1028 05:18:40.915971 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.916231 kubelet[2750]: E1028 05:18:40.916197 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.916231 kubelet[2750]: W1028 05:18:40.916207 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.916499 kubelet[2750]: E1028 05:18:40.916386 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.916956 kubelet[2750]: E1028 05:18:40.916913 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.916956 kubelet[2750]: W1028 05:18:40.916925 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.917121 kubelet[2750]: E1028 05:18:40.917085 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.917559 kubelet[2750]: E1028 05:18:40.917483 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.917559 kubelet[2750]: W1028 05:18:40.917518 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.917700 kubelet[2750]: E1028 05:18:40.917675 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.918010 kubelet[2750]: E1028 05:18:40.917985 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.918010 kubelet[2750]: W1028 05:18:40.917996 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.918185 kubelet[2750]: E1028 05:18:40.918161 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.918573 kubelet[2750]: E1028 05:18:40.918485 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.918573 kubelet[2750]: W1028 05:18:40.918496 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.918573 kubelet[2750]: E1028 05:18:40.918544 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.919098 kubelet[2750]: E1028 05:18:40.919073 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.919098 kubelet[2750]: W1028 05:18:40.919084 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.919354 kubelet[2750]: E1028 05:18:40.919303 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.919618 kubelet[2750]: E1028 05:18:40.919586 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.919618 kubelet[2750]: W1028 05:18:40.919597 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.919758 kubelet[2750]: E1028 05:18:40.919732 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.920065 kubelet[2750]: E1028 05:18:40.920039 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.920065 kubelet[2750]: W1028 05:18:40.920051 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.920204 kubelet[2750]: E1028 05:18:40.920181 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.920581 kubelet[2750]: E1028 05:18:40.920556 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.920581 kubelet[2750]: W1028 05:18:40.920567 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.920771 kubelet[2750]: E1028 05:18:40.920737 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.921039 kubelet[2750]: E1028 05:18:40.921027 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.921094 kubelet[2750]: W1028 05:18:40.921084 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.921213 kubelet[2750]: E1028 05:18:40.921190 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.922067 kubelet[2750]: E1028 05:18:40.922027 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.922067 kubelet[2750]: W1028 05:18:40.922053 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.922598 kubelet[2750]: E1028 05:18:40.922583 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.922927 kubelet[2750]: E1028 05:18:40.922902 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.922927 kubelet[2750]: W1028 05:18:40.922913 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.923061 kubelet[2750]: E1028 05:18:40.923036 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.923385 kubelet[2750]: E1028 05:18:40.923373 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.923472 kubelet[2750]: W1028 05:18:40.923431 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.923602 kubelet[2750]: E1028 05:18:40.923578 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.923852 kubelet[2750]: E1028 05:18:40.923829 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.923852 kubelet[2750]: W1028 05:18:40.923840 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.924047 kubelet[2750]: E1028 05:18:40.924030 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.924187 kubelet[2750]: E1028 05:18:40.924174 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.924263 kubelet[2750]: W1028 05:18:40.924247 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.924455 kubelet[2750]: E1028 05:18:40.924431 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.924778 kubelet[2750]: E1028 05:18:40.924753 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.924778 kubelet[2750]: W1028 05:18:40.924765 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.925124 kubelet[2750]: E1028 05:18:40.924988 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.925472 kubelet[2750]: E1028 05:18:40.925452 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.925532 kubelet[2750]: W1028 05:18:40.925520 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.925624 kubelet[2750]: E1028 05:18:40.925599 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:40.931898 containerd[1617]: time="2025-10-28T05:18:40.931855929Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-9k4vr,Uid:51af4ad7-d74e-45f7-8c48-7dfea57da407,Namespace:calico-system,Attempt:0,} returns sandbox id \"d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2\"" Oct 28 05:18:40.933147 kubelet[2750]: E1028 05:18:40.933103 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:40.933959 kubelet[2750]: E1028 05:18:40.933940 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:40.933959 kubelet[2750]: W1028 05:18:40.933956 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:40.934060 kubelet[2750]: E1028 05:18:40.933970 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:42.246827 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3827784308.mount: Deactivated successfully. Oct 28 05:18:42.666162 containerd[1617]: time="2025-10-28T05:18:42.666085750Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:42.693635 kubelet[2750]: E1028 05:18:42.693583 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:18:42.723066 containerd[1617]: time="2025-10-28T05:18:42.723013185Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=35234628" Oct 28 05:18:42.724548 containerd[1617]: time="2025-10-28T05:18:42.724510571Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:42.726688 containerd[1617]: time="2025-10-28T05:18:42.726662811Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:42.727288 containerd[1617]: time="2025-10-28T05:18:42.727260079Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 1.87256013s" Oct 28 05:18:42.727344 containerd[1617]: time="2025-10-28T05:18:42.727296630Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Oct 28 05:18:42.728262 containerd[1617]: time="2025-10-28T05:18:42.728111264Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Oct 28 05:18:42.736026 containerd[1617]: time="2025-10-28T05:18:42.735987810Z" level=info msg="CreateContainer within sandbox \"95ad9e3a99699b6a2bd9e6ec6e6372ae59281e43eb2943b5652e0fa86b7addd7\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Oct 28 05:18:42.741693 containerd[1617]: time="2025-10-28T05:18:42.741660239Z" level=info msg="Container 13e5babc1977537731a4663ae24a53cfa879a4d5089253f6ace08c74e7afe191: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:42.747747 containerd[1617]: time="2025-10-28T05:18:42.747720762Z" level=info msg="CreateContainer within sandbox \"95ad9e3a99699b6a2bd9e6ec6e6372ae59281e43eb2943b5652e0fa86b7addd7\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"13e5babc1977537731a4663ae24a53cfa879a4d5089253f6ace08c74e7afe191\"" Oct 28 05:18:42.748224 containerd[1617]: time="2025-10-28T05:18:42.748172173Z" level=info msg="StartContainer for \"13e5babc1977537731a4663ae24a53cfa879a4d5089253f6ace08c74e7afe191\"" Oct 28 05:18:42.749306 containerd[1617]: time="2025-10-28T05:18:42.749270185Z" level=info msg="connecting to shim 13e5babc1977537731a4663ae24a53cfa879a4d5089253f6ace08c74e7afe191" address="unix:///run/containerd/s/af225b9610014eb3a33c81ce2479aca51df5422aec6f195c2594f3502b1b379d" protocol=ttrpc version=3 Oct 28 05:18:42.770543 systemd[1]: Started cri-containerd-13e5babc1977537731a4663ae24a53cfa879a4d5089253f6ace08c74e7afe191.scope - libcontainer container 13e5babc1977537731a4663ae24a53cfa879a4d5089253f6ace08c74e7afe191. Oct 28 05:18:42.818983 containerd[1617]: time="2025-10-28T05:18:42.818939433Z" level=info msg="StartContainer for \"13e5babc1977537731a4663ae24a53cfa879a4d5089253f6ace08c74e7afe191\" returns successfully" Oct 28 05:18:43.769383 kubelet[2750]: E1028 05:18:43.769346 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:43.817510 kubelet[2750]: E1028 05:18:43.817472 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.817510 kubelet[2750]: W1028 05:18:43.817495 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.817510 kubelet[2750]: E1028 05:18:43.817514 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.817693 kubelet[2750]: E1028 05:18:43.817686 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.817719 kubelet[2750]: W1028 05:18:43.817694 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.817719 kubelet[2750]: E1028 05:18:43.817702 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.817926 kubelet[2750]: E1028 05:18:43.817903 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.817926 kubelet[2750]: W1028 05:18:43.817913 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.817926 kubelet[2750]: E1028 05:18:43.817920 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.818115 kubelet[2750]: E1028 05:18:43.818094 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.818115 kubelet[2750]: W1028 05:18:43.818104 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.818115 kubelet[2750]: E1028 05:18:43.818112 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.818321 kubelet[2750]: E1028 05:18:43.818301 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.818321 kubelet[2750]: W1028 05:18:43.818310 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.818321 kubelet[2750]: E1028 05:18:43.818318 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.818510 kubelet[2750]: E1028 05:18:43.818498 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.818510 kubelet[2750]: W1028 05:18:43.818507 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.818571 kubelet[2750]: E1028 05:18:43.818515 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.818677 kubelet[2750]: E1028 05:18:43.818666 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.818677 kubelet[2750]: W1028 05:18:43.818674 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.818732 kubelet[2750]: E1028 05:18:43.818682 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.818882 kubelet[2750]: E1028 05:18:43.818866 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.818882 kubelet[2750]: W1028 05:18:43.818882 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.818958 kubelet[2750]: E1028 05:18:43.818894 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.819110 kubelet[2750]: E1028 05:18:43.819096 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.819110 kubelet[2750]: W1028 05:18:43.819106 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.819184 kubelet[2750]: E1028 05:18:43.819116 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.819326 kubelet[2750]: E1028 05:18:43.819296 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.819326 kubelet[2750]: W1028 05:18:43.819307 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.819326 kubelet[2750]: E1028 05:18:43.819315 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.820535 kubelet[2750]: E1028 05:18:43.820498 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.820535 kubelet[2750]: W1028 05:18:43.820524 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.820626 kubelet[2750]: E1028 05:18:43.820553 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.820808 kubelet[2750]: E1028 05:18:43.820791 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.820808 kubelet[2750]: W1028 05:18:43.820802 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.820882 kubelet[2750]: E1028 05:18:43.820811 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.820992 kubelet[2750]: E1028 05:18:43.820967 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.820992 kubelet[2750]: W1028 05:18:43.820987 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.821048 kubelet[2750]: E1028 05:18:43.820994 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.821233 kubelet[2750]: E1028 05:18:43.821213 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.821233 kubelet[2750]: W1028 05:18:43.821227 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.821316 kubelet[2750]: E1028 05:18:43.821239 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.821482 kubelet[2750]: E1028 05:18:43.821465 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.821482 kubelet[2750]: W1028 05:18:43.821477 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.821567 kubelet[2750]: E1028 05:18:43.821485 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.834848 kubelet[2750]: E1028 05:18:43.834818 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.834848 kubelet[2750]: W1028 05:18:43.834834 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.834848 kubelet[2750]: E1028 05:18:43.834848 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.835105 kubelet[2750]: E1028 05:18:43.835080 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.835105 kubelet[2750]: W1028 05:18:43.835093 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.835165 kubelet[2750]: E1028 05:18:43.835108 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.835326 kubelet[2750]: E1028 05:18:43.835301 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.835326 kubelet[2750]: W1028 05:18:43.835318 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.835382 kubelet[2750]: E1028 05:18:43.835340 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.835571 kubelet[2750]: E1028 05:18:43.835558 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.835571 kubelet[2750]: W1028 05:18:43.835568 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.835632 kubelet[2750]: E1028 05:18:43.835581 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.835785 kubelet[2750]: E1028 05:18:43.835767 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.835785 kubelet[2750]: W1028 05:18:43.835776 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.835832 kubelet[2750]: E1028 05:18:43.835790 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.836000 kubelet[2750]: E1028 05:18:43.835981 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.836000 kubelet[2750]: W1028 05:18:43.835993 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.836117 kubelet[2750]: E1028 05:18:43.836007 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.836287 kubelet[2750]: E1028 05:18:43.836269 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.836287 kubelet[2750]: W1028 05:18:43.836283 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.836342 kubelet[2750]: E1028 05:18:43.836299 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.836503 kubelet[2750]: E1028 05:18:43.836490 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.836503 kubelet[2750]: W1028 05:18:43.836500 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.836558 kubelet[2750]: E1028 05:18:43.836514 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.836705 kubelet[2750]: E1028 05:18:43.836688 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.836705 kubelet[2750]: W1028 05:18:43.836702 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.836756 kubelet[2750]: E1028 05:18:43.836718 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.836906 kubelet[2750]: E1028 05:18:43.836891 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.836906 kubelet[2750]: W1028 05:18:43.836902 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.836956 kubelet[2750]: E1028 05:18:43.836915 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.837107 kubelet[2750]: E1028 05:18:43.837093 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.837107 kubelet[2750]: W1028 05:18:43.837103 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.837161 kubelet[2750]: E1028 05:18:43.837117 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.837298 kubelet[2750]: E1028 05:18:43.837282 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.837298 kubelet[2750]: W1028 05:18:43.837293 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.837345 kubelet[2750]: E1028 05:18:43.837306 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.837521 kubelet[2750]: E1028 05:18:43.837505 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.837521 kubelet[2750]: W1028 05:18:43.837517 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.837570 kubelet[2750]: E1028 05:18:43.837529 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.837790 kubelet[2750]: E1028 05:18:43.837765 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.837790 kubelet[2750]: W1028 05:18:43.837779 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.837790 kubelet[2750]: E1028 05:18:43.837788 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.837960 kubelet[2750]: E1028 05:18:43.837944 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.837960 kubelet[2750]: W1028 05:18:43.837956 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.838016 kubelet[2750]: E1028 05:18:43.837966 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.838172 kubelet[2750]: E1028 05:18:43.838157 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.838172 kubelet[2750]: W1028 05:18:43.838167 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.838223 kubelet[2750]: E1028 05:18:43.838179 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.838367 kubelet[2750]: E1028 05:18:43.838350 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.838367 kubelet[2750]: W1028 05:18:43.838362 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.838431 kubelet[2750]: E1028 05:18:43.838372 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.838557 kubelet[2750]: E1028 05:18:43.838542 2750 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 28 05:18:43.838557 kubelet[2750]: W1028 05:18:43.838553 2750 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 28 05:18:43.838613 kubelet[2750]: E1028 05:18:43.838563 2750 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 28 05:18:43.984246 containerd[1617]: time="2025-10-28T05:18:43.984191460Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:43.985019 containerd[1617]: time="2025-10-28T05:18:43.984962340Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=4446754" Oct 28 05:18:43.986279 containerd[1617]: time="2025-10-28T05:18:43.986238017Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:43.988805 containerd[1617]: time="2025-10-28T05:18:43.988771871Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:43.989444 containerd[1617]: time="2025-10-28T05:18:43.989389009Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 1.261249577s" Oct 28 05:18:43.989500 containerd[1617]: time="2025-10-28T05:18:43.989446697Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Oct 28 05:18:43.991234 containerd[1617]: time="2025-10-28T05:18:43.991198422Z" level=info msg="CreateContainer within sandbox \"d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Oct 28 05:18:43.999797 containerd[1617]: time="2025-10-28T05:18:43.999759895Z" level=info msg="Container aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:44.008862 containerd[1617]: time="2025-10-28T05:18:44.008813526Z" level=info msg="CreateContainer within sandbox \"d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f\"" Oct 28 05:18:44.010413 containerd[1617]: time="2025-10-28T05:18:44.009417058Z" level=info msg="StartContainer for \"aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f\"" Oct 28 05:18:44.013067 containerd[1617]: time="2025-10-28T05:18:44.013006435Z" level=info msg="connecting to shim aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f" address="unix:///run/containerd/s/ff54ef066755ecfaefbc3f0f98152dc00bfd2bc80aab441ce2153e9cb809ef44" protocol=ttrpc version=3 Oct 28 05:18:44.037686 systemd[1]: Started cri-containerd-aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f.scope - libcontainer container aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f. Oct 28 05:18:44.125637 containerd[1617]: time="2025-10-28T05:18:44.125551124Z" level=info msg="StartContainer for \"aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f\" returns successfully" Oct 28 05:18:44.139852 systemd[1]: cri-containerd-aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f.scope: Deactivated successfully. Oct 28 05:18:44.142670 containerd[1617]: time="2025-10-28T05:18:44.142437615Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f\" id:\"aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f\" pid:3465 exited_at:{seconds:1761628724 nanos:141945934}" Oct 28 05:18:44.142670 containerd[1617]: time="2025-10-28T05:18:44.142497898Z" level=info msg="received exit event container_id:\"aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f\" id:\"aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f\" pid:3465 exited_at:{seconds:1761628724 nanos:141945934}" Oct 28 05:18:44.167524 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-aa774563ef367b43634366879461b12dbc715d32cfd761e50ea6317bf0f1a36f-rootfs.mount: Deactivated successfully. Oct 28 05:18:44.692935 kubelet[2750]: E1028 05:18:44.692884 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:18:44.773514 kubelet[2750]: I1028 05:18:44.773468 2750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 28 05:18:44.774074 kubelet[2750]: E1028 05:18:44.773840 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:44.774074 kubelet[2750]: E1028 05:18:44.774015 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:44.774883 containerd[1617]: time="2025-10-28T05:18:44.774613754Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Oct 28 05:18:44.787116 kubelet[2750]: I1028 05:18:44.787045 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-64bb95c58d-v8r42" podStartSLOduration=2.912305268 podStartE2EDuration="4.787027096s" podCreationTimestamp="2025-10-28 05:18:40 +0000 UTC" firstStartedPulling="2025-10-28 05:18:40.853209292 +0000 UTC m=+20.252192263" lastFinishedPulling="2025-10-28 05:18:42.72793112 +0000 UTC m=+22.126914091" observedRunningTime="2025-10-28 05:18:43.781627679 +0000 UTC m=+23.180610650" watchObservedRunningTime="2025-10-28 05:18:44.787027096 +0000 UTC m=+24.186010067" Oct 28 05:18:46.693123 kubelet[2750]: E1028 05:18:46.693070 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:18:47.291159 containerd[1617]: time="2025-10-28T05:18:47.291100614Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:47.291942 containerd[1617]: time="2025-10-28T05:18:47.291885148Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70446859" Oct 28 05:18:47.293007 containerd[1617]: time="2025-10-28T05:18:47.292969393Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:47.295001 containerd[1617]: time="2025-10-28T05:18:47.294940420Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:47.295480 containerd[1617]: time="2025-10-28T05:18:47.295454294Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 2.520806493s" Oct 28 05:18:47.295519 containerd[1617]: time="2025-10-28T05:18:47.295483304Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Oct 28 05:18:47.297688 containerd[1617]: time="2025-10-28T05:18:47.297660951Z" level=info msg="CreateContainer within sandbox \"d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Oct 28 05:18:47.306226 containerd[1617]: time="2025-10-28T05:18:47.306181837Z" level=info msg="Container 27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:47.313996 containerd[1617]: time="2025-10-28T05:18:47.313948752Z" level=info msg="CreateContainer within sandbox \"d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b\"" Oct 28 05:18:47.314574 containerd[1617]: time="2025-10-28T05:18:47.314539199Z" level=info msg="StartContainer for \"27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b\"" Oct 28 05:18:47.315970 containerd[1617]: time="2025-10-28T05:18:47.315942349Z" level=info msg="connecting to shim 27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b" address="unix:///run/containerd/s/ff54ef066755ecfaefbc3f0f98152dc00bfd2bc80aab441ce2153e9cb809ef44" protocol=ttrpc version=3 Oct 28 05:18:47.337545 systemd[1]: Started cri-containerd-27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b.scope - libcontainer container 27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b. Oct 28 05:18:47.394686 containerd[1617]: time="2025-10-28T05:18:47.394632656Z" level=info msg="StartContainer for \"27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b\" returns successfully" Oct 28 05:18:47.781617 kubelet[2750]: E1028 05:18:47.781568 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:48.483269 systemd[1]: cri-containerd-27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b.scope: Deactivated successfully. Oct 28 05:18:48.483792 systemd[1]: cri-containerd-27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b.scope: Consumed 712ms CPU time, 181.6M memory peak, 4M read from disk, 171.3M written to disk. Oct 28 05:18:48.484955 containerd[1617]: time="2025-10-28T05:18:48.484353353Z" level=info msg="received exit event container_id:\"27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b\" id:\"27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b\" pid:3524 exited_at:{seconds:1761628728 nanos:484098208}" Oct 28 05:18:48.484955 containerd[1617]: time="2025-10-28T05:18:48.484776583Z" level=info msg="TaskExit event in podsandbox handler container_id:\"27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b\" id:\"27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b\" pid:3524 exited_at:{seconds:1761628728 nanos:484098208}" Oct 28 05:18:48.511625 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-27f6c2d16d1b08d37697e6b8afb6e5fcf151d03579120757cabfd60b7d10c14b-rootfs.mount: Deactivated successfully. Oct 28 05:18:48.533086 kubelet[2750]: I1028 05:18:48.533049 2750 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Oct 28 05:18:48.563275 systemd[1]: Created slice kubepods-besteffort-pode5ac4e44_6733_40d1_9793_83c29769783f.slice - libcontainer container kubepods-besteffort-pode5ac4e44_6733_40d1_9793_83c29769783f.slice. Oct 28 05:18:48.572546 systemd[1]: Created slice kubepods-burstable-podcd7d412f_ea6e_4e1b_b8aa_f004a51a7675.slice - libcontainer container kubepods-burstable-podcd7d412f_ea6e_4e1b_b8aa_f004a51a7675.slice. Oct 28 05:18:48.579893 systemd[1]: Created slice kubepods-burstable-pod68f02234_c261_477b_a240_b486ef5a3bfc.slice - libcontainer container kubepods-burstable-pod68f02234_c261_477b_a240_b486ef5a3bfc.slice. Oct 28 05:18:48.587158 systemd[1]: Created slice kubepods-besteffort-pod855b8ecd_afaf_480c_b1b2_51b12d85eb28.slice - libcontainer container kubepods-besteffort-pod855b8ecd_afaf_480c_b1b2_51b12d85eb28.slice. Oct 28 05:18:48.593571 systemd[1]: Created slice kubepods-besteffort-podf075f92a_8aac_4f64_b8e3_246d2217434f.slice - libcontainer container kubepods-besteffort-podf075f92a_8aac_4f64_b8e3_246d2217434f.slice. Oct 28 05:18:48.599056 systemd[1]: Created slice kubepods-besteffort-podfc255c62_41f7_4c05_87d5_240f00279ed4.slice - libcontainer container kubepods-besteffort-podfc255c62_41f7_4c05_87d5_240f00279ed4.slice. Oct 28 05:18:48.606069 systemd[1]: Created slice kubepods-besteffort-pod1c878ea8_6d56_4615_a865_dddfefeda8f6.slice - libcontainer container kubepods-besteffort-pod1c878ea8_6d56_4615_a865_dddfefeda8f6.slice. Oct 28 05:18:48.666886 kubelet[2750]: I1028 05:18:48.666820 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e5ac4e44-6733-40d1-9793-83c29769783f-whisker-backend-key-pair\") pod \"whisker-5b598bc676-cjf5s\" (UID: \"e5ac4e44-6733-40d1-9793-83c29769783f\") " pod="calico-system/whisker-5b598bc676-cjf5s" Oct 28 05:18:48.666886 kubelet[2750]: I1028 05:18:48.666874 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5ac4e44-6733-40d1-9793-83c29769783f-whisker-ca-bundle\") pod \"whisker-5b598bc676-cjf5s\" (UID: \"e5ac4e44-6733-40d1-9793-83c29769783f\") " pod="calico-system/whisker-5b598bc676-cjf5s" Oct 28 05:18:48.666886 kubelet[2750]: I1028 05:18:48.666890 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djdgc\" (UniqueName: \"kubernetes.io/projected/68f02234-c261-477b-a240-b486ef5a3bfc-kube-api-access-djdgc\") pod \"coredns-668d6bf9bc-wj4pc\" (UID: \"68f02234-c261-477b-a240-b486ef5a3bfc\") " pod="kube-system/coredns-668d6bf9bc-wj4pc" Oct 28 05:18:48.667077 kubelet[2750]: I1028 05:18:48.666908 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/855b8ecd-afaf-480c-b1b2-51b12d85eb28-goldmane-key-pair\") pod \"goldmane-666569f655-78lhj\" (UID: \"855b8ecd-afaf-480c-b1b2-51b12d85eb28\") " pod="calico-system/goldmane-666569f655-78lhj" Oct 28 05:18:48.667077 kubelet[2750]: I1028 05:18:48.666930 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc255c62-41f7-4c05-87d5-240f00279ed4-tigera-ca-bundle\") pod \"calico-kube-controllers-64d5cb6b45-29rc8\" (UID: \"fc255c62-41f7-4c05-87d5-240f00279ed4\") " pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" Oct 28 05:18:48.667077 kubelet[2750]: I1028 05:18:48.666946 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbnx6\" (UniqueName: \"kubernetes.io/projected/e5ac4e44-6733-40d1-9793-83c29769783f-kube-api-access-kbnx6\") pod \"whisker-5b598bc676-cjf5s\" (UID: \"e5ac4e44-6733-40d1-9793-83c29769783f\") " pod="calico-system/whisker-5b598bc676-cjf5s" Oct 28 05:18:48.667077 kubelet[2750]: I1028 05:18:48.666972 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68f02234-c261-477b-a240-b486ef5a3bfc-config-volume\") pod \"coredns-668d6bf9bc-wj4pc\" (UID: \"68f02234-c261-477b-a240-b486ef5a3bfc\") " pod="kube-system/coredns-668d6bf9bc-wj4pc" Oct 28 05:18:48.667077 kubelet[2750]: I1028 05:18:48.666988 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/855b8ecd-afaf-480c-b1b2-51b12d85eb28-goldmane-ca-bundle\") pod \"goldmane-666569f655-78lhj\" (UID: \"855b8ecd-afaf-480c-b1b2-51b12d85eb28\") " pod="calico-system/goldmane-666569f655-78lhj" Oct 28 05:18:48.667211 kubelet[2750]: I1028 05:18:48.667040 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q95c9\" (UniqueName: \"kubernetes.io/projected/cd7d412f-ea6e-4e1b-b8aa-f004a51a7675-kube-api-access-q95c9\") pod \"coredns-668d6bf9bc-t8j8m\" (UID: \"cd7d412f-ea6e-4e1b-b8aa-f004a51a7675\") " pod="kube-system/coredns-668d6bf9bc-t8j8m" Oct 28 05:18:48.667211 kubelet[2750]: I1028 05:18:48.667084 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/855b8ecd-afaf-480c-b1b2-51b12d85eb28-config\") pod \"goldmane-666569f655-78lhj\" (UID: \"855b8ecd-afaf-480c-b1b2-51b12d85eb28\") " pod="calico-system/goldmane-666569f655-78lhj" Oct 28 05:18:48.667211 kubelet[2750]: I1028 05:18:48.667101 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq9j2\" (UniqueName: \"kubernetes.io/projected/855b8ecd-afaf-480c-b1b2-51b12d85eb28-kube-api-access-mq9j2\") pod \"goldmane-666569f655-78lhj\" (UID: \"855b8ecd-afaf-480c-b1b2-51b12d85eb28\") " pod="calico-system/goldmane-666569f655-78lhj" Oct 28 05:18:48.667211 kubelet[2750]: I1028 05:18:48.667126 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmbbw\" (UniqueName: \"kubernetes.io/projected/f075f92a-8aac-4f64-b8e3-246d2217434f-kube-api-access-kmbbw\") pod \"calico-apiserver-7658475cc6-2hfhm\" (UID: \"f075f92a-8aac-4f64-b8e3-246d2217434f\") " pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" Oct 28 05:18:48.667211 kubelet[2750]: I1028 05:18:48.667144 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd7d412f-ea6e-4e1b-b8aa-f004a51a7675-config-volume\") pod \"coredns-668d6bf9bc-t8j8m\" (UID: \"cd7d412f-ea6e-4e1b-b8aa-f004a51a7675\") " pod="kube-system/coredns-668d6bf9bc-t8j8m" Oct 28 05:18:48.667327 kubelet[2750]: I1028 05:18:48.667201 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mpfx\" (UniqueName: \"kubernetes.io/projected/fc255c62-41f7-4c05-87d5-240f00279ed4-kube-api-access-2mpfx\") pod \"calico-kube-controllers-64d5cb6b45-29rc8\" (UID: \"fc255c62-41f7-4c05-87d5-240f00279ed4\") " pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" Oct 28 05:18:48.667327 kubelet[2750]: I1028 05:18:48.667222 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/1c878ea8-6d56-4615-a865-dddfefeda8f6-calico-apiserver-certs\") pod \"calico-apiserver-7658475cc6-hxjx7\" (UID: \"1c878ea8-6d56-4615-a865-dddfefeda8f6\") " pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" Oct 28 05:18:48.667327 kubelet[2750]: I1028 05:18:48.667247 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77w5b\" (UniqueName: \"kubernetes.io/projected/1c878ea8-6d56-4615-a865-dddfefeda8f6-kube-api-access-77w5b\") pod \"calico-apiserver-7658475cc6-hxjx7\" (UID: \"1c878ea8-6d56-4615-a865-dddfefeda8f6\") " pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" Oct 28 05:18:48.667327 kubelet[2750]: I1028 05:18:48.667265 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f075f92a-8aac-4f64-b8e3-246d2217434f-calico-apiserver-certs\") pod \"calico-apiserver-7658475cc6-2hfhm\" (UID: \"f075f92a-8aac-4f64-b8e3-246d2217434f\") " pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" Oct 28 05:18:48.699135 systemd[1]: Created slice kubepods-besteffort-pod144455de_c836_43b8_85ee_1c567bc354f0.slice - libcontainer container kubepods-besteffort-pod144455de_c836_43b8_85ee_1c567bc354f0.slice. Oct 28 05:18:48.701550 containerd[1617]: time="2025-10-28T05:18:48.701506295Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-tfkvb,Uid:144455de-c836-43b8-85ee-1c567bc354f0,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:48.802568 kubelet[2750]: E1028 05:18:48.801635 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:48.803664 containerd[1617]: time="2025-10-28T05:18:48.803614813Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Oct 28 05:18:48.845605 containerd[1617]: time="2025-10-28T05:18:48.845532398Z" level=error msg="Failed to destroy network for sandbox \"dd0d3a15010fb6d9cf805e485907ed58b563d8311155a096d4c1cc2fd8d3834c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:48.846785 containerd[1617]: time="2025-10-28T05:18:48.846746409Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-tfkvb,Uid:144455de-c836-43b8-85ee-1c567bc354f0,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"dd0d3a15010fb6d9cf805e485907ed58b563d8311155a096d4c1cc2fd8d3834c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:48.855688 kubelet[2750]: E1028 05:18:48.855638 2750 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dd0d3a15010fb6d9cf805e485907ed58b563d8311155a096d4c1cc2fd8d3834c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:48.855853 kubelet[2750]: E1028 05:18:48.855715 2750 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dd0d3a15010fb6d9cf805e485907ed58b563d8311155a096d4c1cc2fd8d3834c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-tfkvb" Oct 28 05:18:48.855853 kubelet[2750]: E1028 05:18:48.855737 2750 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dd0d3a15010fb6d9cf805e485907ed58b563d8311155a096d4c1cc2fd8d3834c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-tfkvb" Oct 28 05:18:48.855853 kubelet[2750]: E1028 05:18:48.855785 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-tfkvb_calico-system(144455de-c836-43b8-85ee-1c567bc354f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-tfkvb_calico-system(144455de-c836-43b8-85ee-1c567bc354f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dd0d3a15010fb6d9cf805e485907ed58b563d8311155a096d4c1cc2fd8d3834c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:18:48.869432 containerd[1617]: time="2025-10-28T05:18:48.868683700Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5b598bc676-cjf5s,Uid:e5ac4e44-6733-40d1-9793-83c29769783f,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:48.877060 kubelet[2750]: E1028 05:18:48.877010 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:48.877896 containerd[1617]: time="2025-10-28T05:18:48.877824543Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-t8j8m,Uid:cd7d412f-ea6e-4e1b-b8aa-f004a51a7675,Namespace:kube-system,Attempt:0,}" Oct 28 05:18:48.884506 kubelet[2750]: E1028 05:18:48.884468 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:48.884982 containerd[1617]: time="2025-10-28T05:18:48.884942320Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wj4pc,Uid:68f02234-c261-477b-a240-b486ef5a3bfc,Namespace:kube-system,Attempt:0,}" Oct 28 05:18:48.890240 containerd[1617]: time="2025-10-28T05:18:48.890183068Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-78lhj,Uid:855b8ecd-afaf-480c-b1b2-51b12d85eb28,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:48.897423 containerd[1617]: time="2025-10-28T05:18:48.897257309Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7658475cc6-2hfhm,Uid:f075f92a-8aac-4f64-b8e3-246d2217434f,Namespace:calico-apiserver,Attempt:0,}" Oct 28 05:18:48.902238 containerd[1617]: time="2025-10-28T05:18:48.902217075Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-64d5cb6b45-29rc8,Uid:fc255c62-41f7-4c05-87d5-240f00279ed4,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:48.910002 containerd[1617]: time="2025-10-28T05:18:48.909238834Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7658475cc6-hxjx7,Uid:1c878ea8-6d56-4615-a865-dddfefeda8f6,Namespace:calico-apiserver,Attempt:0,}" Oct 28 05:18:48.994581 containerd[1617]: time="2025-10-28T05:18:48.994518959Z" level=error msg="Failed to destroy network for sandbox \"ac75f9b685b6a4c9aab456d5d38dbf2c9881b67fff727aee22cfcc935c6b512d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:48.996583 containerd[1617]: time="2025-10-28T05:18:48.996525745Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5b598bc676-cjf5s,Uid:e5ac4e44-6733-40d1-9793-83c29769783f,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac75f9b685b6a4c9aab456d5d38dbf2c9881b67fff727aee22cfcc935c6b512d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:48.996888 kubelet[2750]: E1028 05:18:48.996800 2750 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac75f9b685b6a4c9aab456d5d38dbf2c9881b67fff727aee22cfcc935c6b512d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:48.996994 kubelet[2750]: E1028 05:18:48.996916 2750 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac75f9b685b6a4c9aab456d5d38dbf2c9881b67fff727aee22cfcc935c6b512d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5b598bc676-cjf5s" Oct 28 05:18:48.996994 kubelet[2750]: E1028 05:18:48.996950 2750 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac75f9b685b6a4c9aab456d5d38dbf2c9881b67fff727aee22cfcc935c6b512d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5b598bc676-cjf5s" Oct 28 05:18:48.997047 kubelet[2750]: E1028 05:18:48.997003 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-5b598bc676-cjf5s_calico-system(e5ac4e44-6733-40d1-9793-83c29769783f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-5b598bc676-cjf5s_calico-system(e5ac4e44-6733-40d1-9793-83c29769783f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ac75f9b685b6a4c9aab456d5d38dbf2c9881b67fff727aee22cfcc935c6b512d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-5b598bc676-cjf5s" podUID="e5ac4e44-6733-40d1-9793-83c29769783f" Oct 28 05:18:49.052962 containerd[1617]: time="2025-10-28T05:18:49.030904934Z" level=error msg="Failed to destroy network for sandbox \"ee221bc3fb8d7cce22006fe89c10646355b70ed679a2dca99b1103532a912077\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.053546 containerd[1617]: time="2025-10-28T05:18:49.042597658Z" level=error msg="Failed to destroy network for sandbox \"5d3b0bfd87693da86b627c4bbe52d076898f40de979eaf95754f9f91e608486c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.053758 containerd[1617]: time="2025-10-28T05:18:49.044506819Z" level=error msg="Failed to destroy network for sandbox \"83ecbf1d064514e8eb8bf86ec1f035df8bd8a50bf745f7bfe223cafe1bbcd64c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.054158 containerd[1617]: time="2025-10-28T05:18:49.042881176Z" level=error msg="Failed to destroy network for sandbox \"84aa388d5d37abf97cf418e4aef96270d6619d57c27b79cbbc77b9a8ce6ee523\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.055888 containerd[1617]: time="2025-10-28T05:18:49.055765452Z" level=error msg="Failed to destroy network for sandbox \"440e5ee192e8164157eb140d393a1d3d020b61c65ada1617a0fee15cda253ea9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.056523 containerd[1617]: time="2025-10-28T05:18:49.056481629Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7658475cc6-hxjx7,Uid:1c878ea8-6d56-4615-a865-dddfefeda8f6,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee221bc3fb8d7cce22006fe89c10646355b70ed679a2dca99b1103532a912077\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.056790 kubelet[2750]: E1028 05:18:49.056739 2750 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee221bc3fb8d7cce22006fe89c10646355b70ed679a2dca99b1103532a912077\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.056849 kubelet[2750]: E1028 05:18:49.056821 2750 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee221bc3fb8d7cce22006fe89c10646355b70ed679a2dca99b1103532a912077\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" Oct 28 05:18:49.056878 kubelet[2750]: E1028 05:18:49.056848 2750 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee221bc3fb8d7cce22006fe89c10646355b70ed679a2dca99b1103532a912077\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" Oct 28 05:18:49.056941 kubelet[2750]: E1028 05:18:49.056907 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7658475cc6-hxjx7_calico-apiserver(1c878ea8-6d56-4615-a865-dddfefeda8f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7658475cc6-hxjx7_calico-apiserver(1c878ea8-6d56-4615-a865-dddfefeda8f6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ee221bc3fb8d7cce22006fe89c10646355b70ed679a2dca99b1103532a912077\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" podUID="1c878ea8-6d56-4615-a865-dddfefeda8f6" Oct 28 05:18:49.057914 containerd[1617]: time="2025-10-28T05:18:49.057823234Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-78lhj,Uid:855b8ecd-afaf-480c-b1b2-51b12d85eb28,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"83ecbf1d064514e8eb8bf86ec1f035df8bd8a50bf745f7bfe223cafe1bbcd64c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.058108 kubelet[2750]: E1028 05:18:49.058067 2750 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"83ecbf1d064514e8eb8bf86ec1f035df8bd8a50bf745f7bfe223cafe1bbcd64c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.058163 kubelet[2750]: E1028 05:18:49.058135 2750 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"83ecbf1d064514e8eb8bf86ec1f035df8bd8a50bf745f7bfe223cafe1bbcd64c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-78lhj" Oct 28 05:18:49.058188 kubelet[2750]: E1028 05:18:49.058159 2750 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"83ecbf1d064514e8eb8bf86ec1f035df8bd8a50bf745f7bfe223cafe1bbcd64c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-78lhj" Oct 28 05:18:49.058243 kubelet[2750]: E1028 05:18:49.058204 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-78lhj_calico-system(855b8ecd-afaf-480c-b1b2-51b12d85eb28)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-78lhj_calico-system(855b8ecd-afaf-480c-b1b2-51b12d85eb28)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"83ecbf1d064514e8eb8bf86ec1f035df8bd8a50bf745f7bfe223cafe1bbcd64c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-78lhj" podUID="855b8ecd-afaf-480c-b1b2-51b12d85eb28" Oct 28 05:18:49.058631 containerd[1617]: time="2025-10-28T05:18:49.058585532Z" level=error msg="Failed to destroy network for sandbox \"457a5e6a883dc07e41fefdf8d5f063925b8c2def82b4642bb1d52835f72b5e46\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.058979 containerd[1617]: time="2025-10-28T05:18:49.058924767Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wj4pc,Uid:68f02234-c261-477b-a240-b486ef5a3bfc,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d3b0bfd87693da86b627c4bbe52d076898f40de979eaf95754f9f91e608486c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.059180 kubelet[2750]: E1028 05:18:49.059089 2750 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d3b0bfd87693da86b627c4bbe52d076898f40de979eaf95754f9f91e608486c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.059180 kubelet[2750]: E1028 05:18:49.059115 2750 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d3b0bfd87693da86b627c4bbe52d076898f40de979eaf95754f9f91e608486c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-wj4pc" Oct 28 05:18:49.059180 kubelet[2750]: E1028 05:18:49.059129 2750 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d3b0bfd87693da86b627c4bbe52d076898f40de979eaf95754f9f91e608486c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-wj4pc" Oct 28 05:18:49.059294 kubelet[2750]: E1028 05:18:49.059162 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-wj4pc_kube-system(68f02234-c261-477b-a240-b486ef5a3bfc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-wj4pc_kube-system(68f02234-c261-477b-a240-b486ef5a3bfc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5d3b0bfd87693da86b627c4bbe52d076898f40de979eaf95754f9f91e608486c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-wj4pc" podUID="68f02234-c261-477b-a240-b486ef5a3bfc" Oct 28 05:18:49.059982 containerd[1617]: time="2025-10-28T05:18:49.059918221Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-64d5cb6b45-29rc8,Uid:fc255c62-41f7-4c05-87d5-240f00279ed4,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"84aa388d5d37abf97cf418e4aef96270d6619d57c27b79cbbc77b9a8ce6ee523\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.060240 kubelet[2750]: E1028 05:18:49.060195 2750 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84aa388d5d37abf97cf418e4aef96270d6619d57c27b79cbbc77b9a8ce6ee523\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.060324 kubelet[2750]: E1028 05:18:49.060291 2750 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84aa388d5d37abf97cf418e4aef96270d6619d57c27b79cbbc77b9a8ce6ee523\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" Oct 28 05:18:49.060368 kubelet[2750]: E1028 05:18:49.060323 2750 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84aa388d5d37abf97cf418e4aef96270d6619d57c27b79cbbc77b9a8ce6ee523\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" Oct 28 05:18:49.060508 kubelet[2750]: E1028 05:18:49.060373 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-64d5cb6b45-29rc8_calico-system(fc255c62-41f7-4c05-87d5-240f00279ed4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-64d5cb6b45-29rc8_calico-system(fc255c62-41f7-4c05-87d5-240f00279ed4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"84aa388d5d37abf97cf418e4aef96270d6619d57c27b79cbbc77b9a8ce6ee523\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" podUID="fc255c62-41f7-4c05-87d5-240f00279ed4" Oct 28 05:18:49.060967 containerd[1617]: time="2025-10-28T05:18:49.060891660Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-t8j8m,Uid:cd7d412f-ea6e-4e1b-b8aa-f004a51a7675,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"440e5ee192e8164157eb140d393a1d3d020b61c65ada1617a0fee15cda253ea9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.061191 kubelet[2750]: E1028 05:18:49.061154 2750 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"440e5ee192e8164157eb140d393a1d3d020b61c65ada1617a0fee15cda253ea9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.061261 kubelet[2750]: E1028 05:18:49.061207 2750 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"440e5ee192e8164157eb140d393a1d3d020b61c65ada1617a0fee15cda253ea9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-t8j8m" Oct 28 05:18:49.061261 kubelet[2750]: E1028 05:18:49.061234 2750 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"440e5ee192e8164157eb140d393a1d3d020b61c65ada1617a0fee15cda253ea9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-t8j8m" Oct 28 05:18:49.061316 kubelet[2750]: E1028 05:18:49.061274 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-t8j8m_kube-system(cd7d412f-ea6e-4e1b-b8aa-f004a51a7675)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-t8j8m_kube-system(cd7d412f-ea6e-4e1b-b8aa-f004a51a7675)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"440e5ee192e8164157eb140d393a1d3d020b61c65ada1617a0fee15cda253ea9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-t8j8m" podUID="cd7d412f-ea6e-4e1b-b8aa-f004a51a7675" Oct 28 05:18:49.061953 containerd[1617]: time="2025-10-28T05:18:49.061910349Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7658475cc6-2hfhm,Uid:f075f92a-8aac-4f64-b8e3-246d2217434f,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"457a5e6a883dc07e41fefdf8d5f063925b8c2def82b4642bb1d52835f72b5e46\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.062136 kubelet[2750]: E1028 05:18:49.062106 2750 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"457a5e6a883dc07e41fefdf8d5f063925b8c2def82b4642bb1d52835f72b5e46\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 28 05:18:49.062191 kubelet[2750]: E1028 05:18:49.062154 2750 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"457a5e6a883dc07e41fefdf8d5f063925b8c2def82b4642bb1d52835f72b5e46\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" Oct 28 05:18:49.062191 kubelet[2750]: E1028 05:18:49.062176 2750 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"457a5e6a883dc07e41fefdf8d5f063925b8c2def82b4642bb1d52835f72b5e46\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" Oct 28 05:18:49.062254 kubelet[2750]: E1028 05:18:49.062224 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7658475cc6-2hfhm_calico-apiserver(f075f92a-8aac-4f64-b8e3-246d2217434f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7658475cc6-2hfhm_calico-apiserver(f075f92a-8aac-4f64-b8e3-246d2217434f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"457a5e6a883dc07e41fefdf8d5f063925b8c2def82b4642bb1d52835f72b5e46\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" podUID="f075f92a-8aac-4f64-b8e3-246d2217434f" Oct 28 05:18:49.517506 systemd[1]: run-netns-cni\x2d6cb1c86b\x2d3cd2\x2d8de4\x2d4cf8\x2d2b34b078b281.mount: Deactivated successfully. Oct 28 05:18:55.280196 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2391022220.mount: Deactivated successfully. Oct 28 05:18:55.882848 containerd[1617]: time="2025-10-28T05:18:55.882784646Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:55.883636 containerd[1617]: time="2025-10-28T05:18:55.883572439Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156883675" Oct 28 05:18:55.884949 containerd[1617]: time="2025-10-28T05:18:55.884911227Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:55.886934 containerd[1617]: time="2025-10-28T05:18:55.886886792Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 28 05:18:55.887344 containerd[1617]: time="2025-10-28T05:18:55.887311026Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 7.083640375s" Oct 28 05:18:55.887344 containerd[1617]: time="2025-10-28T05:18:55.887339039Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Oct 28 05:18:55.898793 containerd[1617]: time="2025-10-28T05:18:55.898749499Z" level=info msg="CreateContainer within sandbox \"d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Oct 28 05:18:55.915259 containerd[1617]: time="2025-10-28T05:18:55.915187750Z" level=info msg="Container 538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:18:56.200372 containerd[1617]: time="2025-10-28T05:18:56.200197654Z" level=info msg="CreateContainer within sandbox \"d63583b309f1f14a3dac725d9c1520dfae52958fe1bb930e7663a0e8172d09a2\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef\"" Oct 28 05:18:56.202424 containerd[1617]: time="2025-10-28T05:18:56.201020683Z" level=info msg="StartContainer for \"538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef\"" Oct 28 05:18:56.202923 containerd[1617]: time="2025-10-28T05:18:56.202894364Z" level=info msg="connecting to shim 538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef" address="unix:///run/containerd/s/ff54ef066755ecfaefbc3f0f98152dc00bfd2bc80aab441ce2153e9cb809ef44" protocol=ttrpc version=3 Oct 28 05:18:56.244575 systemd[1]: Started cri-containerd-538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef.scope - libcontainer container 538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef. Oct 28 05:18:56.292677 containerd[1617]: time="2025-10-28T05:18:56.292639448Z" level=info msg="StartContainer for \"538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef\" returns successfully" Oct 28 05:18:56.371259 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Oct 28 05:18:56.372245 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Oct 28 05:18:56.520056 kubelet[2750]: I1028 05:18:56.519912 2750 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e5ac4e44-6733-40d1-9793-83c29769783f-whisker-backend-key-pair\") pod \"e5ac4e44-6733-40d1-9793-83c29769783f\" (UID: \"e5ac4e44-6733-40d1-9793-83c29769783f\") " Oct 28 05:18:56.520056 kubelet[2750]: I1028 05:18:56.519978 2750 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbnx6\" (UniqueName: \"kubernetes.io/projected/e5ac4e44-6733-40d1-9793-83c29769783f-kube-api-access-kbnx6\") pod \"e5ac4e44-6733-40d1-9793-83c29769783f\" (UID: \"e5ac4e44-6733-40d1-9793-83c29769783f\") " Oct 28 05:18:56.520056 kubelet[2750]: I1028 05:18:56.520014 2750 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5ac4e44-6733-40d1-9793-83c29769783f-whisker-ca-bundle\") pod \"e5ac4e44-6733-40d1-9793-83c29769783f\" (UID: \"e5ac4e44-6733-40d1-9793-83c29769783f\") " Oct 28 05:18:56.522585 kubelet[2750]: I1028 05:18:56.522545 2750 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5ac4e44-6733-40d1-9793-83c29769783f-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "e5ac4e44-6733-40d1-9793-83c29769783f" (UID: "e5ac4e44-6733-40d1-9793-83c29769783f"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Oct 28 05:18:56.526303 systemd[1]: var-lib-kubelet-pods-e5ac4e44\x2d6733\x2d40d1\x2d9793\x2d83c29769783f-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkbnx6.mount: Deactivated successfully. Oct 28 05:18:56.527199 kubelet[2750]: I1028 05:18:56.527156 2750 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5ac4e44-6733-40d1-9793-83c29769783f-kube-api-access-kbnx6" (OuterVolumeSpecName: "kube-api-access-kbnx6") pod "e5ac4e44-6733-40d1-9793-83c29769783f" (UID: "e5ac4e44-6733-40d1-9793-83c29769783f"). InnerVolumeSpecName "kube-api-access-kbnx6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Oct 28 05:18:56.527510 kubelet[2750]: I1028 05:18:56.527451 2750 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ac4e44-6733-40d1-9793-83c29769783f-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "e5ac4e44-6733-40d1-9793-83c29769783f" (UID: "e5ac4e44-6733-40d1-9793-83c29769783f"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Oct 28 05:18:56.531045 systemd[1]: var-lib-kubelet-pods-e5ac4e44\x2d6733\x2d40d1\x2d9793\x2d83c29769783f-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Oct 28 05:18:56.622190 kubelet[2750]: I1028 05:18:56.622140 2750 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5ac4e44-6733-40d1-9793-83c29769783f-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Oct 28 05:18:56.622190 kubelet[2750]: I1028 05:18:56.622176 2750 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e5ac4e44-6733-40d1-9793-83c29769783f-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Oct 28 05:18:56.622190 kubelet[2750]: I1028 05:18:56.622185 2750 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kbnx6\" (UniqueName: \"kubernetes.io/projected/e5ac4e44-6733-40d1-9793-83c29769783f-kube-api-access-kbnx6\") on node \"localhost\" DevicePath \"\"" Oct 28 05:18:56.701046 systemd[1]: Removed slice kubepods-besteffort-pode5ac4e44_6733_40d1_9793_83c29769783f.slice - libcontainer container kubepods-besteffort-pode5ac4e44_6733_40d1_9793_83c29769783f.slice. Oct 28 05:18:56.830138 kubelet[2750]: E1028 05:18:56.830105 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:56.922103 kubelet[2750]: I1028 05:18:56.921581 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-9k4vr" podStartSLOduration=1.967419971 podStartE2EDuration="16.921563214s" podCreationTimestamp="2025-10-28 05:18:40 +0000 UTC" firstStartedPulling="2025-10-28 05:18:40.934038323 +0000 UTC m=+20.333021295" lastFinishedPulling="2025-10-28 05:18:55.888181557 +0000 UTC m=+35.287164538" observedRunningTime="2025-10-28 05:18:56.921010567 +0000 UTC m=+36.319993528" watchObservedRunningTime="2025-10-28 05:18:56.921563214 +0000 UTC m=+36.320546175" Oct 28 05:18:56.936706 systemd[1]: Created slice kubepods-besteffort-podc69e2212_1789_4672_9606_362def9455d7.slice - libcontainer container kubepods-besteffort-podc69e2212_1789_4672_9606_362def9455d7.slice. Oct 28 05:18:57.025387 kubelet[2750]: I1028 05:18:57.025336 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/c69e2212-1789-4672-9606-362def9455d7-whisker-backend-key-pair\") pod \"whisker-57bb759ddb-zpxmt\" (UID: \"c69e2212-1789-4672-9606-362def9455d7\") " pod="calico-system/whisker-57bb759ddb-zpxmt" Oct 28 05:18:57.025387 kubelet[2750]: I1028 05:18:57.025381 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c69e2212-1789-4672-9606-362def9455d7-whisker-ca-bundle\") pod \"whisker-57bb759ddb-zpxmt\" (UID: \"c69e2212-1789-4672-9606-362def9455d7\") " pod="calico-system/whisker-57bb759ddb-zpxmt" Oct 28 05:18:57.025387 kubelet[2750]: I1028 05:18:57.025416 2750 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hr24\" (UniqueName: \"kubernetes.io/projected/c69e2212-1789-4672-9606-362def9455d7-kube-api-access-7hr24\") pod \"whisker-57bb759ddb-zpxmt\" (UID: \"c69e2212-1789-4672-9606-362def9455d7\") " pod="calico-system/whisker-57bb759ddb-zpxmt" Oct 28 05:18:57.243504 containerd[1617]: time="2025-10-28T05:18:57.243360361Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-57bb759ddb-zpxmt,Uid:c69e2212-1789-4672-9606-362def9455d7,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:57.404408 systemd-networkd[1517]: calid46ae545e00: Link UP Oct 28 05:18:57.404688 systemd-networkd[1517]: calid46ae545e00: Gained carrier Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.277 [INFO][3908] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.295 [INFO][3908] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--57bb759ddb--zpxmt-eth0 whisker-57bb759ddb- calico-system c69e2212-1789-4672-9606-362def9455d7 946 0 2025-10-28 05:18:56 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:57bb759ddb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-57bb759ddb-zpxmt eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calid46ae545e00 [] [] }} ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Namespace="calico-system" Pod="whisker-57bb759ddb-zpxmt" WorkloadEndpoint="localhost-k8s-whisker--57bb759ddb--zpxmt-" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.295 [INFO][3908] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Namespace="calico-system" Pod="whisker-57bb759ddb-zpxmt" WorkloadEndpoint="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.361 [INFO][3922] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" HandleID="k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Workload="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.362 [INFO][3922] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" HandleID="k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Workload="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000c9710), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-57bb759ddb-zpxmt", "timestamp":"2025-10-28 05:18:57.3616739 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.362 [INFO][3922] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.362 [INFO][3922] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.362 [INFO][3922] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.370 [INFO][3922] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.377 [INFO][3922] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.380 [INFO][3922] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.382 [INFO][3922] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.383 [INFO][3922] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.383 [INFO][3922] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.385 [INFO][3922] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0 Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.388 [INFO][3922] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.393 [INFO][3922] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.393 [INFO][3922] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" host="localhost" Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.393 [INFO][3922] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 28 05:18:57.416369 containerd[1617]: 2025-10-28 05:18:57.393 [INFO][3922] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" HandleID="k8s-pod-network.70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Workload="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" Oct 28 05:18:57.417017 containerd[1617]: 2025-10-28 05:18:57.397 [INFO][3908] cni-plugin/k8s.go 418: Populated endpoint ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Namespace="calico-system" Pod="whisker-57bb759ddb-zpxmt" WorkloadEndpoint="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--57bb759ddb--zpxmt-eth0", GenerateName:"whisker-57bb759ddb-", Namespace:"calico-system", SelfLink:"", UID:"c69e2212-1789-4672-9606-362def9455d7", ResourceVersion:"946", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"57bb759ddb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-57bb759ddb-zpxmt", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calid46ae545e00", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:18:57.417017 containerd[1617]: 2025-10-28 05:18:57.397 [INFO][3908] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Namespace="calico-system" Pod="whisker-57bb759ddb-zpxmt" WorkloadEndpoint="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" Oct 28 05:18:57.417017 containerd[1617]: 2025-10-28 05:18:57.397 [INFO][3908] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid46ae545e00 ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Namespace="calico-system" Pod="whisker-57bb759ddb-zpxmt" WorkloadEndpoint="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" Oct 28 05:18:57.417017 containerd[1617]: 2025-10-28 05:18:57.404 [INFO][3908] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Namespace="calico-system" Pod="whisker-57bb759ddb-zpxmt" WorkloadEndpoint="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" Oct 28 05:18:57.417017 containerd[1617]: 2025-10-28 05:18:57.405 [INFO][3908] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Namespace="calico-system" Pod="whisker-57bb759ddb-zpxmt" WorkloadEndpoint="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--57bb759ddb--zpxmt-eth0", GenerateName:"whisker-57bb759ddb-", Namespace:"calico-system", SelfLink:"", UID:"c69e2212-1789-4672-9606-362def9455d7", ResourceVersion:"946", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"57bb759ddb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0", Pod:"whisker-57bb759ddb-zpxmt", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calid46ae545e00", MAC:"56:e2:58:e2:a8:3f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:18:57.417017 containerd[1617]: 2025-10-28 05:18:57.413 [INFO][3908] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" Namespace="calico-system" Pod="whisker-57bb759ddb-zpxmt" WorkloadEndpoint="localhost-k8s-whisker--57bb759ddb--zpxmt-eth0" Oct 28 05:18:57.472494 containerd[1617]: time="2025-10-28T05:18:57.472424930Z" level=info msg="connecting to shim 70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0" address="unix:///run/containerd/s/2dd3a82d8b224b429cd4b96e36ab368bf4ebc3d67778981ea7b183608649defa" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:57.501572 systemd[1]: Started cri-containerd-70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0.scope - libcontainer container 70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0. Oct 28 05:18:57.519667 systemd-resolved[1366]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 28 05:18:57.818421 containerd[1617]: time="2025-10-28T05:18:57.817453236Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-57bb759ddb-zpxmt,Uid:c69e2212-1789-4672-9606-362def9455d7,Namespace:calico-system,Attempt:0,} returns sandbox id \"70d48f1580735dbbaa45ef12e4e084247b62ded3c053724dabc7436d0bc79ed0\"" Oct 28 05:18:57.823415 containerd[1617]: time="2025-10-28T05:18:57.823345780Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 28 05:18:57.832812 kubelet[2750]: E1028 05:18:57.832770 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:57.977757 containerd[1617]: time="2025-10-28T05:18:57.977701971Z" level=info msg="TaskExit event in podsandbox handler container_id:\"538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef\" id:\"3ad8bd136531075df09fa098fe8c3ead2637907a74ceb66584dec4e506255f05\" pid:4096 exit_status:1 exited_at:{seconds:1761628737 nanos:977310229}" Oct 28 05:18:58.206510 containerd[1617]: time="2025-10-28T05:18:58.206325055Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:18:58.207756 containerd[1617]: time="2025-10-28T05:18:58.207692254Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 28 05:18:58.212766 containerd[1617]: time="2025-10-28T05:18:58.212712723Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 28 05:18:58.213040 kubelet[2750]: E1028 05:18:58.212985 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 28 05:18:58.213114 kubelet[2750]: E1028 05:18:58.213055 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 28 05:18:58.221697 kubelet[2750]: E1028 05:18:58.221624 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:2f3301ac9dce4d0092d2b57652ea2b6c,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7hr24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-57bb759ddb-zpxmt_calico-system(c69e2212-1789-4672-9606-362def9455d7): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 28 05:18:58.224764 containerd[1617]: time="2025-10-28T05:18:58.223528789Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 28 05:18:58.695368 kubelet[2750]: I1028 05:18:58.695318 2750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5ac4e44-6733-40d1-9793-83c29769783f" path="/var/lib/kubelet/pods/e5ac4e44-6733-40d1-9793-83c29769783f/volumes" Oct 28 05:18:58.762409 containerd[1617]: time="2025-10-28T05:18:58.762342221Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:18:58.763710 containerd[1617]: time="2025-10-28T05:18:58.763657182Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 28 05:18:58.763710 containerd[1617]: time="2025-10-28T05:18:58.763746230Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 28 05:18:58.764155 kubelet[2750]: E1028 05:18:58.764029 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 28 05:18:58.764155 kubelet[2750]: E1028 05:18:58.764115 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 28 05:18:58.764534 kubelet[2750]: E1028 05:18:58.764472 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7hr24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-57bb759ddb-zpxmt_calico-system(c69e2212-1789-4672-9606-362def9455d7): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 28 05:18:58.765738 kubelet[2750]: E1028 05:18:58.765682 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-57bb759ddb-zpxmt" podUID="c69e2212-1789-4672-9606-362def9455d7" Oct 28 05:18:58.835749 kubelet[2750]: E1028 05:18:58.835391 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:18:58.844271 kubelet[2750]: E1028 05:18:58.843752 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-57bb759ddb-zpxmt" podUID="c69e2212-1789-4672-9606-362def9455d7" Oct 28 05:18:58.938945 containerd[1617]: time="2025-10-28T05:18:58.938900473Z" level=info msg="TaskExit event in podsandbox handler container_id:\"538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef\" id:\"4c3dd6dceaa0c4cda656addfa2ec602b4b175bf4fcfc550e62b49341c319bde4\" pid:4145 exit_status:1 exited_at:{seconds:1761628738 nanos:938591307}" Oct 28 05:18:58.960630 systemd-networkd[1517]: calid46ae545e00: Gained IPv6LL Oct 28 05:18:59.693836 containerd[1617]: time="2025-10-28T05:18:59.693774448Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-78lhj,Uid:855b8ecd-afaf-480c-b1b2-51b12d85eb28,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:59.694059 containerd[1617]: time="2025-10-28T05:18:59.694024050Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-64d5cb6b45-29rc8,Uid:fc255c62-41f7-4c05-87d5-240f00279ed4,Namespace:calico-system,Attempt:0,}" Oct 28 05:18:59.794660 systemd-networkd[1517]: cali95db480253c: Link UP Oct 28 05:18:59.795868 systemd-networkd[1517]: cali95db480253c: Gained carrier Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.725 [INFO][4174] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.735 [INFO][4174] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0 calico-kube-controllers-64d5cb6b45- calico-system fc255c62-41f7-4c05-87d5-240f00279ed4 864 0 2025-10-28 05:18:40 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:64d5cb6b45 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-64d5cb6b45-29rc8 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali95db480253c [] [] }} ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Namespace="calico-system" Pod="calico-kube-controllers-64d5cb6b45-29rc8" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.735 [INFO][4174] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Namespace="calico-system" Pod="calico-kube-controllers-64d5cb6b45-29rc8" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.760 [INFO][4190] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" HandleID="k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Workload="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.760 [INFO][4190] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" HandleID="k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Workload="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002e7560), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-64d5cb6b45-29rc8", "timestamp":"2025-10-28 05:18:59.760592277 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.760 [INFO][4190] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.762 [INFO][4190] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.762 [INFO][4190] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.768 [INFO][4190] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.772 [INFO][4190] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.776 [INFO][4190] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.778 [INFO][4190] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.779 [INFO][4190] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.779 [INFO][4190] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.781 [INFO][4190] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0 Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.784 [INFO][4190] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.789 [INFO][4190] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.789 [INFO][4190] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" host="localhost" Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.789 [INFO][4190] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 28 05:18:59.811637 containerd[1617]: 2025-10-28 05:18:59.789 [INFO][4190] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" HandleID="k8s-pod-network.7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Workload="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" Oct 28 05:18:59.812532 containerd[1617]: 2025-10-28 05:18:59.792 [INFO][4174] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Namespace="calico-system" Pod="calico-kube-controllers-64d5cb6b45-29rc8" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0", GenerateName:"calico-kube-controllers-64d5cb6b45-", Namespace:"calico-system", SelfLink:"", UID:"fc255c62-41f7-4c05-87d5-240f00279ed4", ResourceVersion:"864", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"64d5cb6b45", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-64d5cb6b45-29rc8", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali95db480253c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:18:59.812532 containerd[1617]: 2025-10-28 05:18:59.792 [INFO][4174] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Namespace="calico-system" Pod="calico-kube-controllers-64d5cb6b45-29rc8" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" Oct 28 05:18:59.812532 containerd[1617]: 2025-10-28 05:18:59.792 [INFO][4174] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali95db480253c ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Namespace="calico-system" Pod="calico-kube-controllers-64d5cb6b45-29rc8" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" Oct 28 05:18:59.812532 containerd[1617]: 2025-10-28 05:18:59.797 [INFO][4174] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Namespace="calico-system" Pod="calico-kube-controllers-64d5cb6b45-29rc8" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" Oct 28 05:18:59.812532 containerd[1617]: 2025-10-28 05:18:59.797 [INFO][4174] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Namespace="calico-system" Pod="calico-kube-controllers-64d5cb6b45-29rc8" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0", GenerateName:"calico-kube-controllers-64d5cb6b45-", Namespace:"calico-system", SelfLink:"", UID:"fc255c62-41f7-4c05-87d5-240f00279ed4", ResourceVersion:"864", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"64d5cb6b45", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0", Pod:"calico-kube-controllers-64d5cb6b45-29rc8", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali95db480253c", MAC:"1e:13:65:9f:04:3c", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:18:59.812532 containerd[1617]: 2025-10-28 05:18:59.806 [INFO][4174] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" Namespace="calico-system" Pod="calico-kube-controllers-64d5cb6b45-29rc8" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--64d5cb6b45--29rc8-eth0" Oct 28 05:18:59.838081 containerd[1617]: time="2025-10-28T05:18:59.838023592Z" level=info msg="connecting to shim 7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0" address="unix:///run/containerd/s/8212ab9fa644607a29fb7613558cfa314d4e79bf8c9ff942b3a57a7a3de41b77" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:18:59.873455 systemd[1]: Started cri-containerd-7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0.scope - libcontainer container 7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0. Oct 28 05:18:59.901639 systemd-resolved[1366]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 28 05:18:59.976697 containerd[1617]: time="2025-10-28T05:18:59.976522608Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-64d5cb6b45-29rc8,Uid:fc255c62-41f7-4c05-87d5-240f00279ed4,Namespace:calico-system,Attempt:0,} returns sandbox id \"7885e8be4a5358f8c6d00f1380335a3e0429444cdf52887972c70a0f079f5ac0\"" Oct 28 05:18:59.978855 containerd[1617]: time="2025-10-28T05:18:59.978822442Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 28 05:19:00.000799 systemd-networkd[1517]: cali2050e95425d: Link UP Oct 28 05:19:00.001571 systemd-networkd[1517]: cali2050e95425d: Gained carrier Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.723 [INFO][4162] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.736 [INFO][4162] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--666569f655--78lhj-eth0 goldmane-666569f655- calico-system 855b8ecd-afaf-480c-b1b2-51b12d85eb28 866 0 2025-10-28 05:18:38 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-666569f655-78lhj eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali2050e95425d [] [] }} ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Namespace="calico-system" Pod="goldmane-666569f655-78lhj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--78lhj-" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.736 [INFO][4162] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Namespace="calico-system" Pod="goldmane-666569f655-78lhj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--78lhj-eth0" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.762 [INFO][4192] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" HandleID="k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Workload="localhost-k8s-goldmane--666569f655--78lhj-eth0" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.762 [INFO][4192] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" HandleID="k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Workload="localhost-k8s-goldmane--666569f655--78lhj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e6b0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-666569f655-78lhj", "timestamp":"2025-10-28 05:18:59.762836446 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.763 [INFO][4192] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.789 [INFO][4192] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.789 [INFO][4192] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.871 [INFO][4192] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.976 [INFO][4192] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.982 [INFO][4192] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.983 [INFO][4192] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.985 [INFO][4192] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.985 [INFO][4192] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.987 [INFO][4192] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.990 [INFO][4192] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.995 [INFO][4192] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.995 [INFO][4192] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" host="localhost" Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.995 [INFO][4192] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 28 05:19:00.016536 containerd[1617]: 2025-10-28 05:18:59.995 [INFO][4192] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" HandleID="k8s-pod-network.4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Workload="localhost-k8s-goldmane--666569f655--78lhj-eth0" Oct 28 05:19:00.017131 containerd[1617]: 2025-10-28 05:18:59.999 [INFO][4162] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Namespace="calico-system" Pod="goldmane-666569f655-78lhj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--78lhj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--78lhj-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"855b8ecd-afaf-480c-b1b2-51b12d85eb28", ResourceVersion:"866", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-666569f655-78lhj", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali2050e95425d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:00.017131 containerd[1617]: 2025-10-28 05:18:59.999 [INFO][4162] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Namespace="calico-system" Pod="goldmane-666569f655-78lhj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--78lhj-eth0" Oct 28 05:19:00.017131 containerd[1617]: 2025-10-28 05:18:59.999 [INFO][4162] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2050e95425d ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Namespace="calico-system" Pod="goldmane-666569f655-78lhj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--78lhj-eth0" Oct 28 05:19:00.017131 containerd[1617]: 2025-10-28 05:19:00.001 [INFO][4162] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Namespace="calico-system" Pod="goldmane-666569f655-78lhj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--78lhj-eth0" Oct 28 05:19:00.017131 containerd[1617]: 2025-10-28 05:19:00.001 [INFO][4162] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Namespace="calico-system" Pod="goldmane-666569f655-78lhj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--78lhj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--78lhj-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"855b8ecd-afaf-480c-b1b2-51b12d85eb28", ResourceVersion:"866", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e", Pod:"goldmane-666569f655-78lhj", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali2050e95425d", MAC:"82:d1:7a:2e:ab:18", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:00.017131 containerd[1617]: 2025-10-28 05:19:00.012 [INFO][4162] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" Namespace="calico-system" Pod="goldmane-666569f655-78lhj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--78lhj-eth0" Oct 28 05:19:00.038920 containerd[1617]: time="2025-10-28T05:19:00.038861767Z" level=info msg="connecting to shim 4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e" address="unix:///run/containerd/s/8c208384606a9157956c935322e8ba94f5626ae81c1e231208ee20ba1f26887a" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:19:00.052201 systemd[1]: Started sshd@7-10.0.0.103:22-10.0.0.1:43830.service - OpenSSH per-connection server daemon (10.0.0.1:43830). Oct 28 05:19:00.079377 systemd[1]: Started cri-containerd-4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e.scope - libcontainer container 4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e. Oct 28 05:19:00.096550 systemd-resolved[1366]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 28 05:19:00.132013 containerd[1617]: time="2025-10-28T05:19:00.131965682Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-78lhj,Uid:855b8ecd-afaf-480c-b1b2-51b12d85eb28,Namespace:calico-system,Attempt:0,} returns sandbox id \"4e50f421657448377010b4616706a5ab7a87060783e45701c0c647ad1368f45e\"" Oct 28 05:19:00.136036 sshd[4314]: Accepted publickey for core from 10.0.0.1 port 43830 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:00.137908 sshd-session[4314]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:00.143043 systemd-logind[1594]: New session 8 of user core. Oct 28 05:19:00.151536 systemd[1]: Started session-8.scope - Session 8 of User core. Oct 28 05:19:00.258442 sshd[4337]: Connection closed by 10.0.0.1 port 43830 Oct 28 05:19:00.258647 sshd-session[4314]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:00.263585 systemd[1]: sshd@7-10.0.0.103:22-10.0.0.1:43830.service: Deactivated successfully. Oct 28 05:19:00.266088 systemd[1]: session-8.scope: Deactivated successfully. Oct 28 05:19:00.267061 systemd-logind[1594]: Session 8 logged out. Waiting for processes to exit. Oct 28 05:19:00.268741 systemd-logind[1594]: Removed session 8. Oct 28 05:19:00.351319 containerd[1617]: time="2025-10-28T05:19:00.351247762Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:00.354461 containerd[1617]: time="2025-10-28T05:19:00.354422051Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 28 05:19:00.354583 containerd[1617]: time="2025-10-28T05:19:00.354475372Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 28 05:19:00.354684 kubelet[2750]: E1028 05:19:00.354630 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 28 05:19:00.354684 kubelet[2750]: E1028 05:19:00.354679 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 28 05:19:00.355583 kubelet[2750]: E1028 05:19:00.354910 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2mpfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-64d5cb6b45-29rc8_calico-system(fc255c62-41f7-4c05-87d5-240f00279ed4): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:00.355681 containerd[1617]: time="2025-10-28T05:19:00.354948527Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 28 05:19:00.356358 kubelet[2750]: E1028 05:19:00.356316 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" podUID="fc255c62-41f7-4c05-87d5-240f00279ed4" Oct 28 05:19:00.682474 containerd[1617]: time="2025-10-28T05:19:00.682409187Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:00.684356 containerd[1617]: time="2025-10-28T05:19:00.684238429Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 28 05:19:00.684356 containerd[1617]: time="2025-10-28T05:19:00.684286961Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:00.684539 kubelet[2750]: E1028 05:19:00.684464 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 28 05:19:00.684539 kubelet[2750]: E1028 05:19:00.684508 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 28 05:19:00.684686 kubelet[2750]: E1028 05:19:00.684617 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mq9j2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-78lhj_calico-system(855b8ecd-afaf-480c-b1b2-51b12d85eb28): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:00.685809 kubelet[2750]: E1028 05:19:00.685766 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-78lhj" podUID="855b8ecd-afaf-480c-b1b2-51b12d85eb28" Oct 28 05:19:00.693683 kubelet[2750]: E1028 05:19:00.693642 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:00.694023 containerd[1617]: time="2025-10-28T05:19:00.693990139Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wj4pc,Uid:68f02234-c261-477b-a240-b486ef5a3bfc,Namespace:kube-system,Attempt:0,}" Oct 28 05:19:00.798469 systemd-networkd[1517]: caliceaf668fd45: Link UP Oct 28 05:19:00.798689 systemd-networkd[1517]: caliceaf668fd45: Gained carrier Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.720 [INFO][4356] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.730 [INFO][4356] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0 coredns-668d6bf9bc- kube-system 68f02234-c261-477b-a240-b486ef5a3bfc 868 0 2025-10-28 05:18:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-wj4pc eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] caliceaf668fd45 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Namespace="kube-system" Pod="coredns-668d6bf9bc-wj4pc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--wj4pc-" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.730 [INFO][4356] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Namespace="kube-system" Pod="coredns-668d6bf9bc-wj4pc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.763 [INFO][4369] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" HandleID="k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Workload="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.763 [INFO][4369] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" HandleID="k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Workload="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e1d0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-wj4pc", "timestamp":"2025-10-28 05:19:00.763282338 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.763 [INFO][4369] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.763 [INFO][4369] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.763 [INFO][4369] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.769 [INFO][4369] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.775 [INFO][4369] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.779 [INFO][4369] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.781 [INFO][4369] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.782 [INFO][4369] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.782 [INFO][4369] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.784 [INFO][4369] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827 Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.787 [INFO][4369] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.792 [INFO][4369] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.792 [INFO][4369] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" host="localhost" Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.792 [INFO][4369] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 28 05:19:00.818979 containerd[1617]: 2025-10-28 05:19:00.792 [INFO][4369] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" HandleID="k8s-pod-network.8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Workload="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" Oct 28 05:19:00.820642 containerd[1617]: 2025-10-28 05:19:00.796 [INFO][4356] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Namespace="kube-system" Pod="coredns-668d6bf9bc-wj4pc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"68f02234-c261-477b-a240-b486ef5a3bfc", ResourceVersion:"868", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-wj4pc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliceaf668fd45", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:00.820642 containerd[1617]: 2025-10-28 05:19:00.796 [INFO][4356] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Namespace="kube-system" Pod="coredns-668d6bf9bc-wj4pc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" Oct 28 05:19:00.820642 containerd[1617]: 2025-10-28 05:19:00.796 [INFO][4356] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliceaf668fd45 ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Namespace="kube-system" Pod="coredns-668d6bf9bc-wj4pc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" Oct 28 05:19:00.820642 containerd[1617]: 2025-10-28 05:19:00.799 [INFO][4356] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Namespace="kube-system" Pod="coredns-668d6bf9bc-wj4pc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" Oct 28 05:19:00.820642 containerd[1617]: 2025-10-28 05:19:00.799 [INFO][4356] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Namespace="kube-system" Pod="coredns-668d6bf9bc-wj4pc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"68f02234-c261-477b-a240-b486ef5a3bfc", ResourceVersion:"868", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827", Pod:"coredns-668d6bf9bc-wj4pc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliceaf668fd45", MAC:"06:6d:d7:7b:d8:ed", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:00.820902 containerd[1617]: 2025-10-28 05:19:00.813 [INFO][4356] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" Namespace="kube-system" Pod="coredns-668d6bf9bc-wj4pc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--wj4pc-eth0" Oct 28 05:19:00.842303 kubelet[2750]: E1028 05:19:00.841438 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" podUID="fc255c62-41f7-4c05-87d5-240f00279ed4" Oct 28 05:19:00.842303 kubelet[2750]: E1028 05:19:00.841446 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-78lhj" podUID="855b8ecd-afaf-480c-b1b2-51b12d85eb28" Oct 28 05:19:00.860591 containerd[1617]: time="2025-10-28T05:19:00.860516722Z" level=info msg="connecting to shim 8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827" address="unix:///run/containerd/s/2612de3bd818000970af0c68a2e73ffc73cb5e7763107190931ac3aa9da364e2" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:19:00.890576 systemd[1]: Started cri-containerd-8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827.scope - libcontainer container 8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827. Oct 28 05:19:00.906000 systemd-resolved[1366]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 28 05:19:00.936908 containerd[1617]: time="2025-10-28T05:19:00.936800126Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wj4pc,Uid:68f02234-c261-477b-a240-b486ef5a3bfc,Namespace:kube-system,Attempt:0,} returns sandbox id \"8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827\"" Oct 28 05:19:00.938349 kubelet[2750]: E1028 05:19:00.938322 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:00.940344 containerd[1617]: time="2025-10-28T05:19:00.940313036Z" level=info msg="CreateContainer within sandbox \"8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 28 05:19:00.986835 containerd[1617]: time="2025-10-28T05:19:00.986768156Z" level=info msg="Container 8d7891e470b286fc17c7f228cf730c5687dde18605b975c224688e82dd82d670: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:19:00.996480 containerd[1617]: time="2025-10-28T05:19:00.996290160Z" level=info msg="CreateContainer within sandbox \"8f195235026fee93e1576ea59f2822d4f04849c5380789f844f2f7483a87c827\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8d7891e470b286fc17c7f228cf730c5687dde18605b975c224688e82dd82d670\"" Oct 28 05:19:00.997121 containerd[1617]: time="2025-10-28T05:19:00.997091868Z" level=info msg="StartContainer for \"8d7891e470b286fc17c7f228cf730c5687dde18605b975c224688e82dd82d670\"" Oct 28 05:19:00.999132 containerd[1617]: time="2025-10-28T05:19:00.999103526Z" level=info msg="connecting to shim 8d7891e470b286fc17c7f228cf730c5687dde18605b975c224688e82dd82d670" address="unix:///run/containerd/s/2612de3bd818000970af0c68a2e73ffc73cb5e7763107190931ac3aa9da364e2" protocol=ttrpc version=3 Oct 28 05:19:01.023635 systemd[1]: Started cri-containerd-8d7891e470b286fc17c7f228cf730c5687dde18605b975c224688e82dd82d670.scope - libcontainer container 8d7891e470b286fc17c7f228cf730c5687dde18605b975c224688e82dd82d670. Oct 28 05:19:01.063579 containerd[1617]: time="2025-10-28T05:19:01.063541116Z" level=info msg="StartContainer for \"8d7891e470b286fc17c7f228cf730c5687dde18605b975c224688e82dd82d670\" returns successfully" Oct 28 05:19:01.136634 systemd-networkd[1517]: cali2050e95425d: Gained IPv6LL Oct 28 05:19:01.200585 systemd-networkd[1517]: cali95db480253c: Gained IPv6LL Oct 28 05:19:01.693724 kubelet[2750]: E1028 05:19:01.693677 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:01.694123 containerd[1617]: time="2025-10-28T05:19:01.694053587Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-t8j8m,Uid:cd7d412f-ea6e-4e1b-b8aa-f004a51a7675,Namespace:kube-system,Attempt:0,}" Oct 28 05:19:01.830731 systemd-networkd[1517]: cali9166cbf5a52: Link UP Oct 28 05:19:01.831582 systemd-networkd[1517]: cali9166cbf5a52: Gained carrier Oct 28 05:19:01.844919 kubelet[2750]: E1028 05:19:01.844895 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:01.846225 kubelet[2750]: E1028 05:19:01.846028 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" podUID="fc255c62-41f7-4c05-87d5-240f00279ed4" Oct 28 05:19:01.846319 kubelet[2750]: E1028 05:19:01.846293 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-78lhj" podUID="855b8ecd-afaf-480c-b1b2-51b12d85eb28" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.736 [INFO][4490] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.746 [INFO][4490] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0 coredns-668d6bf9bc- kube-system cd7d412f-ea6e-4e1b-b8aa-f004a51a7675 867 0 2025-10-28 05:18:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-t8j8m eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali9166cbf5a52 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Namespace="kube-system" Pod="coredns-668d6bf9bc-t8j8m" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--t8j8m-" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.746 [INFO][4490] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Namespace="kube-system" Pod="coredns-668d6bf9bc-t8j8m" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.772 [INFO][4505] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" HandleID="k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Workload="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.773 [INFO][4505] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" HandleID="k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Workload="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002c6fd0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-t8j8m", "timestamp":"2025-10-28 05:19:01.772741054 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.773 [INFO][4505] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.773 [INFO][4505] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.773 [INFO][4505] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.779 [INFO][4505] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.784 [INFO][4505] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.789 [INFO][4505] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.791 [INFO][4505] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.792 [INFO][4505] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.793 [INFO][4505] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.794 [INFO][4505] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0 Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.809 [INFO][4505] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.823 [INFO][4505] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.823 [INFO][4505] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" host="localhost" Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.823 [INFO][4505] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 28 05:19:01.853388 containerd[1617]: 2025-10-28 05:19:01.823 [INFO][4505] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" HandleID="k8s-pod-network.565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Workload="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" Oct 28 05:19:01.854213 containerd[1617]: 2025-10-28 05:19:01.827 [INFO][4490] cni-plugin/k8s.go 418: Populated endpoint ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Namespace="kube-system" Pod="coredns-668d6bf9bc-t8j8m" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"cd7d412f-ea6e-4e1b-b8aa-f004a51a7675", ResourceVersion:"867", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-t8j8m", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali9166cbf5a52", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:01.854213 containerd[1617]: 2025-10-28 05:19:01.828 [INFO][4490] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Namespace="kube-system" Pod="coredns-668d6bf9bc-t8j8m" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" Oct 28 05:19:01.854213 containerd[1617]: 2025-10-28 05:19:01.828 [INFO][4490] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9166cbf5a52 ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Namespace="kube-system" Pod="coredns-668d6bf9bc-t8j8m" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" Oct 28 05:19:01.854213 containerd[1617]: 2025-10-28 05:19:01.831 [INFO][4490] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Namespace="kube-system" Pod="coredns-668d6bf9bc-t8j8m" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" Oct 28 05:19:01.854213 containerd[1617]: 2025-10-28 05:19:01.831 [INFO][4490] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Namespace="kube-system" Pod="coredns-668d6bf9bc-t8j8m" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"cd7d412f-ea6e-4e1b-b8aa-f004a51a7675", ResourceVersion:"867", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0", Pod:"coredns-668d6bf9bc-t8j8m", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali9166cbf5a52", MAC:"5a:f1:0c:b6:62:8d", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:01.854509 containerd[1617]: 2025-10-28 05:19:01.848 [INFO][4490] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" Namespace="kube-system" Pod="coredns-668d6bf9bc-t8j8m" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--t8j8m-eth0" Oct 28 05:19:01.877050 kubelet[2750]: I1028 05:19:01.876988 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-wj4pc" podStartSLOduration=33.876968986 podStartE2EDuration="33.876968986s" podCreationTimestamp="2025-10-28 05:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-28 05:19:01.876059615 +0000 UTC m=+41.275042606" watchObservedRunningTime="2025-10-28 05:19:01.876968986 +0000 UTC m=+41.275951957" Oct 28 05:19:01.934723 containerd[1617]: time="2025-10-28T05:19:01.933744906Z" level=info msg="connecting to shim 565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0" address="unix:///run/containerd/s/39428297da3103f4158a65fde21ffea445a22126c6a15d81490a46e83b757b31" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:19:01.971549 systemd[1]: Started cri-containerd-565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0.scope - libcontainer container 565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0. Oct 28 05:19:01.987691 systemd-resolved[1366]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 28 05:19:02.031088 containerd[1617]: time="2025-10-28T05:19:02.031028772Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-t8j8m,Uid:cd7d412f-ea6e-4e1b-b8aa-f004a51a7675,Namespace:kube-system,Attempt:0,} returns sandbox id \"565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0\"" Oct 28 05:19:02.031882 kubelet[2750]: E1028 05:19:02.031858 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:02.033873 containerd[1617]: time="2025-10-28T05:19:02.033834601Z" level=info msg="CreateContainer within sandbox \"565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 28 05:19:02.104186 containerd[1617]: time="2025-10-28T05:19:02.103562859Z" level=info msg="Container 57ae2afd1992232b86fc9c3f64e618c95a48fa00786466352d6157ebb2f1d3e5: CDI devices from CRI Config.CDIDevices: []" Oct 28 05:19:02.116948 containerd[1617]: time="2025-10-28T05:19:02.116865509Z" level=info msg="CreateContainer within sandbox \"565c7ad6c1361a4784ecb89aca23ecf20bc1775d7df21c3241420547ec1906a0\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"57ae2afd1992232b86fc9c3f64e618c95a48fa00786466352d6157ebb2f1d3e5\"" Oct 28 05:19:02.118416 containerd[1617]: time="2025-10-28T05:19:02.118050861Z" level=info msg="StartContainer for \"57ae2afd1992232b86fc9c3f64e618c95a48fa00786466352d6157ebb2f1d3e5\"" Oct 28 05:19:02.118881 containerd[1617]: time="2025-10-28T05:19:02.118853120Z" level=info msg="connecting to shim 57ae2afd1992232b86fc9c3f64e618c95a48fa00786466352d6157ebb2f1d3e5" address="unix:///run/containerd/s/39428297da3103f4158a65fde21ffea445a22126c6a15d81490a46e83b757b31" protocol=ttrpc version=3 Oct 28 05:19:02.143550 systemd[1]: Started cri-containerd-57ae2afd1992232b86fc9c3f64e618c95a48fa00786466352d6157ebb2f1d3e5.scope - libcontainer container 57ae2afd1992232b86fc9c3f64e618c95a48fa00786466352d6157ebb2f1d3e5. Oct 28 05:19:02.182694 containerd[1617]: time="2025-10-28T05:19:02.182633625Z" level=info msg="StartContainer for \"57ae2afd1992232b86fc9c3f64e618c95a48fa00786466352d6157ebb2f1d3e5\" returns successfully" Oct 28 05:19:02.693978 containerd[1617]: time="2025-10-28T05:19:02.693929821Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7658475cc6-hxjx7,Uid:1c878ea8-6d56-4615-a865-dddfefeda8f6,Namespace:calico-apiserver,Attempt:0,}" Oct 28 05:19:02.694129 containerd[1617]: time="2025-10-28T05:19:02.693929841Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7658475cc6-2hfhm,Uid:f075f92a-8aac-4f64-b8e3-246d2217434f,Namespace:calico-apiserver,Attempt:0,}" Oct 28 05:19:02.736533 systemd-networkd[1517]: caliceaf668fd45: Gained IPv6LL Oct 28 05:19:02.847967 kubelet[2750]: E1028 05:19:02.847725 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:02.847967 kubelet[2750]: E1028 05:19:02.847764 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:03.226358 systemd-networkd[1517]: cali9a4aca66af4: Link UP Oct 28 05:19:03.227499 systemd-networkd[1517]: cali9a4aca66af4: Gained carrier Oct 28 05:19:03.367583 kubelet[2750]: I1028 05:19:03.367455 2750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-t8j8m" podStartSLOduration=35.367430647 podStartE2EDuration="35.367430647s" podCreationTimestamp="2025-10-28 05:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-28 05:19:02.927502179 +0000 UTC m=+42.326485170" watchObservedRunningTime="2025-10-28 05:19:03.367430647 +0000 UTC m=+42.766413618" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:02.870 [INFO][4626] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.005 [INFO][4626] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0 calico-apiserver-7658475cc6- calico-apiserver 1c878ea8-6d56-4615-a865-dddfefeda8f6 869 0 2025-10-28 05:18:36 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7658475cc6 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7658475cc6-hxjx7 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali9a4aca66af4 [] [] }} ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-hxjx7" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.005 [INFO][4626] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-hxjx7" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.041 [INFO][4656] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" HandleID="k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Workload="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.041 [INFO][4656] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" HandleID="k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Workload="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00042b750), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7658475cc6-hxjx7", "timestamp":"2025-10-28 05:19:03.041450115 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.041 [INFO][4656] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.041 [INFO][4656] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.041 [INFO][4656] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.047 [INFO][4656] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.051 [INFO][4656] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.057 [INFO][4656] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.058 [INFO][4656] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.060 [INFO][4656] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.060 [INFO][4656] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.061 [INFO][4656] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968 Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.079 [INFO][4656] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.209 [INFO][4656] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.209 [INFO][4656] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" host="localhost" Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.209 [INFO][4656] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 28 05:19:03.371070 containerd[1617]: 2025-10-28 05:19:03.209 [INFO][4656] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" HandleID="k8s-pod-network.62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Workload="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" Oct 28 05:19:03.371889 containerd[1617]: 2025-10-28 05:19:03.217 [INFO][4626] cni-plugin/k8s.go 418: Populated endpoint ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-hxjx7" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0", GenerateName:"calico-apiserver-7658475cc6-", Namespace:"calico-apiserver", SelfLink:"", UID:"1c878ea8-6d56-4615-a865-dddfefeda8f6", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7658475cc6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7658475cc6-hxjx7", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9a4aca66af4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:03.371889 containerd[1617]: 2025-10-28 05:19:03.217 [INFO][4626] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-hxjx7" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" Oct 28 05:19:03.371889 containerd[1617]: 2025-10-28 05:19:03.217 [INFO][4626] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9a4aca66af4 ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-hxjx7" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" Oct 28 05:19:03.371889 containerd[1617]: 2025-10-28 05:19:03.228 [INFO][4626] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-hxjx7" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" Oct 28 05:19:03.371889 containerd[1617]: 2025-10-28 05:19:03.230 [INFO][4626] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-hxjx7" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0", GenerateName:"calico-apiserver-7658475cc6-", Namespace:"calico-apiserver", SelfLink:"", UID:"1c878ea8-6d56-4615-a865-dddfefeda8f6", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7658475cc6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968", Pod:"calico-apiserver-7658475cc6-hxjx7", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9a4aca66af4", MAC:"ca:c6:f1:a5:c6:8d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:03.371889 containerd[1617]: 2025-10-28 05:19:03.367 [INFO][4626] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-hxjx7" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--hxjx7-eth0" Oct 28 05:19:03.504578 systemd-networkd[1517]: cali9166cbf5a52: Gained IPv6LL Oct 28 05:19:03.693923 containerd[1617]: time="2025-10-28T05:19:03.693869267Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-tfkvb,Uid:144455de-c836-43b8-85ee-1c567bc354f0,Namespace:calico-system,Attempt:0,}" Oct 28 05:19:03.762141 systemd-networkd[1517]: cali80df907124b: Link UP Oct 28 05:19:03.762940 systemd-networkd[1517]: cali80df907124b: Gained carrier Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:02.930 [INFO][4641] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.011 [INFO][4641] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0 calico-apiserver-7658475cc6- calico-apiserver f075f92a-8aac-4f64-b8e3-246d2217434f 870 0 2025-10-28 05:18:36 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7658475cc6 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7658475cc6-2hfhm eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali80df907124b [] [] }} ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-2hfhm" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.013 [INFO][4641] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-2hfhm" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.041 [INFO][4663] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" HandleID="k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Workload="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.041 [INFO][4663] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" HandleID="k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Workload="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001abe30), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7658475cc6-2hfhm", "timestamp":"2025-10-28 05:19:03.041840845 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.041 [INFO][4663] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.209 [INFO][4663] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.209 [INFO][4663] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.220 [INFO][4663] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.227 [INFO][4663] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.234 [INFO][4663] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.366 [INFO][4663] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.369 [INFO][4663] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.369 [INFO][4663] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.371 [INFO][4663] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7 Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.415 [INFO][4663] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.756 [INFO][4663] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.756 [INFO][4663] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" host="localhost" Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.756 [INFO][4663] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 28 05:19:03.836275 containerd[1617]: 2025-10-28 05:19:03.756 [INFO][4663] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" HandleID="k8s-pod-network.d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Workload="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" Oct 28 05:19:03.836934 containerd[1617]: 2025-10-28 05:19:03.759 [INFO][4641] cni-plugin/k8s.go 418: Populated endpoint ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-2hfhm" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0", GenerateName:"calico-apiserver-7658475cc6-", Namespace:"calico-apiserver", SelfLink:"", UID:"f075f92a-8aac-4f64-b8e3-246d2217434f", ResourceVersion:"870", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7658475cc6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7658475cc6-2hfhm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali80df907124b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:03.836934 containerd[1617]: 2025-10-28 05:19:03.759 [INFO][4641] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-2hfhm" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" Oct 28 05:19:03.836934 containerd[1617]: 2025-10-28 05:19:03.759 [INFO][4641] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali80df907124b ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-2hfhm" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" Oct 28 05:19:03.836934 containerd[1617]: 2025-10-28 05:19:03.763 [INFO][4641] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-2hfhm" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" Oct 28 05:19:03.836934 containerd[1617]: 2025-10-28 05:19:03.763 [INFO][4641] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-2hfhm" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0", GenerateName:"calico-apiserver-7658475cc6-", Namespace:"calico-apiserver", SelfLink:"", UID:"f075f92a-8aac-4f64-b8e3-246d2217434f", ResourceVersion:"870", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7658475cc6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7", Pod:"calico-apiserver-7658475cc6-2hfhm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali80df907124b", MAC:"62:80:ad:e7:9c:a5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:03.836934 containerd[1617]: 2025-10-28 05:19:03.832 [INFO][4641] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" Namespace="calico-apiserver" Pod="calico-apiserver-7658475cc6-2hfhm" WorkloadEndpoint="localhost-k8s-calico--apiserver--7658475cc6--2hfhm-eth0" Oct 28 05:19:03.849530 kubelet[2750]: E1028 05:19:03.849475 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:03.851602 kubelet[2750]: E1028 05:19:03.849800 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:04.196347 containerd[1617]: time="2025-10-28T05:19:04.196292171Z" level=info msg="connecting to shim 62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968" address="unix:///run/containerd/s/03151502fc28aabe877a791d20a8ac588b144e8123759af6fb4de2948398b778" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:19:04.234557 systemd[1]: Started cri-containerd-62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968.scope - libcontainer container 62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968. Oct 28 05:19:04.260589 systemd-resolved[1366]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 28 05:19:04.666601 containerd[1617]: time="2025-10-28T05:19:04.666526648Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7658475cc6-hxjx7,Uid:1c878ea8-6d56-4615-a865-dddfefeda8f6,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"62e9cd060b8a332f821828f3d6c83f2c32d71fcb803b31e1710a439aa9411968\"" Oct 28 05:19:04.668287 containerd[1617]: time="2025-10-28T05:19:04.668241832Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 28 05:19:04.852235 kubelet[2750]: E1028 05:19:04.852203 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:05.273581 systemd[1]: Started sshd@8-10.0.0.103:22-10.0.0.1:43838.service - OpenSSH per-connection server daemon (10.0.0.1:43838). Oct 28 05:19:05.296609 systemd-networkd[1517]: cali9a4aca66af4: Gained IPv6LL Oct 28 05:19:05.297248 systemd-networkd[1517]: cali80df907124b: Gained IPv6LL Oct 28 05:19:05.302367 kubelet[2750]: I1028 05:19:05.302333 2750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 28 05:19:05.302927 kubelet[2750]: E1028 05:19:05.302905 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:05.342450 sshd[4809]: Accepted publickey for core from 10.0.0.1 port 43838 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:05.344007 sshd-session[4809]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:05.348597 systemd-logind[1594]: New session 9 of user core. Oct 28 05:19:05.360109 systemd[1]: Started session-9.scope - Session 9 of User core. Oct 28 05:19:05.405838 containerd[1617]: time="2025-10-28T05:19:05.405780401Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:05.546586 containerd[1617]: time="2025-10-28T05:19:05.546409604Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 28 05:19:05.546586 containerd[1617]: time="2025-10-28T05:19:05.546515463Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:05.546795 kubelet[2750]: E1028 05:19:05.546714 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:05.546795 kubelet[2750]: E1028 05:19:05.546764 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:05.546949 kubelet[2750]: E1028 05:19:05.546902 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-77w5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7658475cc6-hxjx7_calico-apiserver(1c878ea8-6d56-4615-a865-dddfefeda8f6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:05.548093 kubelet[2750]: E1028 05:19:05.548060 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" podUID="1c878ea8-6d56-4615-a865-dddfefeda8f6" Oct 28 05:19:05.656841 containerd[1617]: time="2025-10-28T05:19:05.656776748Z" level=info msg="connecting to shim d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7" address="unix:///run/containerd/s/3f85e64d7ab455d8d9dd277c97177eae7cad0d88fa38d032b5dd9454fdb1ba49" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:19:05.691735 systemd[1]: Started cri-containerd-d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7.scope - libcontainer container d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7. Oct 28 05:19:05.706523 systemd-resolved[1366]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 28 05:19:05.737376 sshd[4812]: Connection closed by 10.0.0.1 port 43838 Oct 28 05:19:05.739506 sshd-session[4809]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:05.747369 systemd[1]: sshd@8-10.0.0.103:22-10.0.0.1:43838.service: Deactivated successfully. Oct 28 05:19:05.752792 systemd[1]: session-9.scope: Deactivated successfully. Oct 28 05:19:05.754585 systemd-logind[1594]: Session 9 logged out. Waiting for processes to exit. Oct 28 05:19:05.756909 systemd-logind[1594]: Removed session 9. Oct 28 05:19:05.854251 kubelet[2750]: E1028 05:19:05.854208 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:05.854755 kubelet[2750]: E1028 05:19:05.854451 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" podUID="1c878ea8-6d56-4615-a865-dddfefeda8f6" Oct 28 05:19:05.855145 kubelet[2750]: E1028 05:19:05.855107 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:05.977690 containerd[1617]: time="2025-10-28T05:19:05.977625607Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7658475cc6-2hfhm,Uid:f075f92a-8aac-4f64-b8e3-246d2217434f,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"d7b501a5bb541ffbc2a19b68b05fcf468bd2f9756daaa784a01104040ff9d6c7\"" Oct 28 05:19:05.980250 containerd[1617]: time="2025-10-28T05:19:05.980120184Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 28 05:19:06.104776 systemd-networkd[1517]: cali4e967a34e14: Link UP Oct 28 05:19:06.104970 systemd-networkd[1517]: cali4e967a34e14: Gained carrier Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:03.881 [INFO][4715] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:04.211 [INFO][4715] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--tfkvb-eth0 csi-node-driver- calico-system 144455de-c836-43b8-85ee-1c567bc354f0 761 0 2025-10-28 05:18:40 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-tfkvb eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali4e967a34e14 [] [] }} ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Namespace="calico-system" Pod="csi-node-driver-tfkvb" WorkloadEndpoint="localhost-k8s-csi--node--driver--tfkvb-" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:04.211 [INFO][4715] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Namespace="calico-system" Pod="csi-node-driver-tfkvb" WorkloadEndpoint="localhost-k8s-csi--node--driver--tfkvb-eth0" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:05.195 [INFO][4800] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" HandleID="k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Workload="localhost-k8s-csi--node--driver--tfkvb-eth0" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:05.195 [INFO][4800] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" HandleID="k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Workload="localhost-k8s-csi--node--driver--tfkvb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f6f0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-tfkvb", "timestamp":"2025-10-28 05:19:05.195649359 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:05.195 [INFO][4800] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:05.195 [INFO][4800] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:05.195 [INFO][4800] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:05.358 [INFO][4800] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:05.975 [INFO][4800] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.013 [INFO][4800] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.038 [INFO][4800] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.044 [INFO][4800] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.044 [INFO][4800] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.051 [INFO][4800] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.078 [INFO][4800] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.097 [INFO][4800] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.097 [INFO][4800] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" host="localhost" Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.097 [INFO][4800] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 28 05:19:06.148198 containerd[1617]: 2025-10-28 05:19:06.097 [INFO][4800] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" HandleID="k8s-pod-network.21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Workload="localhost-k8s-csi--node--driver--tfkvb-eth0" Oct 28 05:19:06.148927 containerd[1617]: 2025-10-28 05:19:06.101 [INFO][4715] cni-plugin/k8s.go 418: Populated endpoint ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Namespace="calico-system" Pod="csi-node-driver-tfkvb" WorkloadEndpoint="localhost-k8s-csi--node--driver--tfkvb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--tfkvb-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"144455de-c836-43b8-85ee-1c567bc354f0", ResourceVersion:"761", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-tfkvb", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali4e967a34e14", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:06.148927 containerd[1617]: 2025-10-28 05:19:06.101 [INFO][4715] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Namespace="calico-system" Pod="csi-node-driver-tfkvb" WorkloadEndpoint="localhost-k8s-csi--node--driver--tfkvb-eth0" Oct 28 05:19:06.148927 containerd[1617]: 2025-10-28 05:19:06.101 [INFO][4715] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4e967a34e14 ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Namespace="calico-system" Pod="csi-node-driver-tfkvb" WorkloadEndpoint="localhost-k8s-csi--node--driver--tfkvb-eth0" Oct 28 05:19:06.148927 containerd[1617]: 2025-10-28 05:19:06.104 [INFO][4715] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Namespace="calico-system" Pod="csi-node-driver-tfkvb" WorkloadEndpoint="localhost-k8s-csi--node--driver--tfkvb-eth0" Oct 28 05:19:06.148927 containerd[1617]: 2025-10-28 05:19:06.106 [INFO][4715] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Namespace="calico-system" Pod="csi-node-driver-tfkvb" WorkloadEndpoint="localhost-k8s-csi--node--driver--tfkvb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--tfkvb-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"144455de-c836-43b8-85ee-1c567bc354f0", ResourceVersion:"761", Generation:0, CreationTimestamp:time.Date(2025, time.October, 28, 5, 18, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e", Pod:"csi-node-driver-tfkvb", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali4e967a34e14", MAC:"ea:f2:42:1e:2b:11", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 28 05:19:06.148927 containerd[1617]: 2025-10-28 05:19:06.144 [INFO][4715] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" Namespace="calico-system" Pod="csi-node-driver-tfkvb" WorkloadEndpoint="localhost-k8s-csi--node--driver--tfkvb-eth0" Oct 28 05:19:06.471619 containerd[1617]: time="2025-10-28T05:19:06.471473929Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:06.510137 containerd[1617]: time="2025-10-28T05:19:06.510092501Z" level=info msg="connecting to shim 21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e" address="unix:///run/containerd/s/8bb2dffa5f4aeb5740c44447d10dac558d80c55d4c8f9bb38a6d2866097b8b91" namespace=k8s.io protocol=ttrpc version=3 Oct 28 05:19:06.528043 containerd[1617]: time="2025-10-28T05:19:06.527985063Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 28 05:19:06.528043 containerd[1617]: time="2025-10-28T05:19:06.528036731Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:06.528275 kubelet[2750]: E1028 05:19:06.528234 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:06.528347 kubelet[2750]: E1028 05:19:06.528295 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:06.530000 kubelet[2750]: E1028 05:19:06.529937 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kmbbw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7658475cc6-2hfhm_calico-apiserver(f075f92a-8aac-4f64-b8e3-246d2217434f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:06.531374 kubelet[2750]: E1028 05:19:06.531294 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" podUID="f075f92a-8aac-4f64-b8e3-246d2217434f" Oct 28 05:19:06.547566 systemd[1]: Started cri-containerd-21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e.scope - libcontainer container 21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e. Oct 28 05:19:06.563448 systemd-resolved[1366]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 28 05:19:06.630529 containerd[1617]: time="2025-10-28T05:19:06.630459273Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-tfkvb,Uid:144455de-c836-43b8-85ee-1c567bc354f0,Namespace:calico-system,Attempt:0,} returns sandbox id \"21e36b3552411015d9eaea4d92002cfebd1bc489e6d4731b51f3559844286c6e\"" Oct 28 05:19:06.635197 containerd[1617]: time="2025-10-28T05:19:06.635116928Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 28 05:19:06.666972 systemd-networkd[1517]: vxlan.calico: Link UP Oct 28 05:19:06.666984 systemd-networkd[1517]: vxlan.calico: Gained carrier Oct 28 05:19:06.860954 kubelet[2750]: E1028 05:19:06.859067 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" podUID="f075f92a-8aac-4f64-b8e3-246d2217434f" Oct 28 05:19:07.017060 containerd[1617]: time="2025-10-28T05:19:07.016966770Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:07.026067 containerd[1617]: time="2025-10-28T05:19:07.026003869Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 28 05:19:07.026209 containerd[1617]: time="2025-10-28T05:19:07.026092177Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 28 05:19:07.026357 kubelet[2750]: E1028 05:19:07.026302 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 28 05:19:07.026447 kubelet[2750]: E1028 05:19:07.026367 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 28 05:19:07.026602 kubelet[2750]: E1028 05:19:07.026521 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-chrf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-tfkvb_calico-system(144455de-c836-43b8-85ee-1c567bc354f0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:07.028661 containerd[1617]: time="2025-10-28T05:19:07.028621346Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 28 05:19:07.413241 containerd[1617]: time="2025-10-28T05:19:07.413179543Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:07.442216 containerd[1617]: time="2025-10-28T05:19:07.442115991Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 28 05:19:07.442453 containerd[1617]: time="2025-10-28T05:19:07.442171746Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 28 05:19:07.442536 kubelet[2750]: E1028 05:19:07.442482 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 28 05:19:07.442593 kubelet[2750]: E1028 05:19:07.442545 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 28 05:19:07.442729 kubelet[2750]: E1028 05:19:07.442664 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-chrf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-tfkvb_calico-system(144455de-c836-43b8-85ee-1c567bc354f0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:07.443946 kubelet[2750]: E1028 05:19:07.443895 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:19:07.536584 systemd-networkd[1517]: cali4e967a34e14: Gained IPv6LL Oct 28 05:19:07.792597 systemd-networkd[1517]: vxlan.calico: Gained IPv6LL Oct 28 05:19:07.861569 kubelet[2750]: E1028 05:19:07.861379 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" podUID="f075f92a-8aac-4f64-b8e3-246d2217434f" Oct 28 05:19:07.861569 kubelet[2750]: E1028 05:19:07.861345 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:19:10.753317 systemd[1]: Started sshd@9-10.0.0.103:22-10.0.0.1:56544.service - OpenSSH per-connection server daemon (10.0.0.1:56544). Oct 28 05:19:10.812429 sshd[5069]: Accepted publickey for core from 10.0.0.1 port 56544 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:10.814225 sshd-session[5069]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:10.821475 systemd-logind[1594]: New session 10 of user core. Oct 28 05:19:10.826645 systemd[1]: Started session-10.scope - Session 10 of User core. Oct 28 05:19:10.914390 sshd[5073]: Connection closed by 10.0.0.1 port 56544 Oct 28 05:19:10.914769 sshd-session[5069]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:10.921386 systemd[1]: sshd@9-10.0.0.103:22-10.0.0.1:56544.service: Deactivated successfully. Oct 28 05:19:10.924305 systemd[1]: session-10.scope: Deactivated successfully. Oct 28 05:19:10.925822 systemd-logind[1594]: Session 10 logged out. Waiting for processes to exit. Oct 28 05:19:10.927140 systemd-logind[1594]: Removed session 10. Oct 28 05:19:11.693784 containerd[1617]: time="2025-10-28T05:19:11.693725184Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 28 05:19:12.013324 containerd[1617]: time="2025-10-28T05:19:12.013193683Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:12.014490 containerd[1617]: time="2025-10-28T05:19:12.014432362Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 28 05:19:12.014628 containerd[1617]: time="2025-10-28T05:19:12.014441018Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 28 05:19:12.014763 kubelet[2750]: E1028 05:19:12.014706 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 28 05:19:12.015093 kubelet[2750]: E1028 05:19:12.014765 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 28 05:19:12.015093 kubelet[2750]: E1028 05:19:12.014902 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:2f3301ac9dce4d0092d2b57652ea2b6c,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7hr24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-57bb759ddb-zpxmt_calico-system(c69e2212-1789-4672-9606-362def9455d7): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:12.017532 containerd[1617]: time="2025-10-28T05:19:12.017507989Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 28 05:19:12.365651 containerd[1617]: time="2025-10-28T05:19:12.365575010Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:12.366811 containerd[1617]: time="2025-10-28T05:19:12.366768163Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 28 05:19:12.366867 containerd[1617]: time="2025-10-28T05:19:12.366820342Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 28 05:19:12.367067 kubelet[2750]: E1028 05:19:12.367004 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 28 05:19:12.367133 kubelet[2750]: E1028 05:19:12.367073 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 28 05:19:12.367272 kubelet[2750]: E1028 05:19:12.367221 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7hr24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-57bb759ddb-zpxmt_calico-system(c69e2212-1789-4672-9606-362def9455d7): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:12.368529 kubelet[2750]: E1028 05:19:12.368468 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-57bb759ddb-zpxmt" podUID="c69e2212-1789-4672-9606-362def9455d7" Oct 28 05:19:15.694535 containerd[1617]: time="2025-10-28T05:19:15.694097276Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 28 05:19:15.928911 systemd[1]: Started sshd@10-10.0.0.103:22-10.0.0.1:56556.service - OpenSSH per-connection server daemon (10.0.0.1:56556). Oct 28 05:19:15.998774 sshd[5096]: Accepted publickey for core from 10.0.0.1 port 56556 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:16.000794 sshd-session[5096]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:16.005498 systemd-logind[1594]: New session 11 of user core. Oct 28 05:19:16.021657 systemd[1]: Started session-11.scope - Session 11 of User core. Oct 28 05:19:16.096137 containerd[1617]: time="2025-10-28T05:19:16.096084162Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:16.213733 sshd[5099]: Connection closed by 10.0.0.1 port 56556 Oct 28 05:19:16.214265 sshd-session[5096]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:16.219345 systemd[1]: sshd@10-10.0.0.103:22-10.0.0.1:56556.service: Deactivated successfully. Oct 28 05:19:16.221718 systemd[1]: session-11.scope: Deactivated successfully. Oct 28 05:19:16.222747 systemd-logind[1594]: Session 11 logged out. Waiting for processes to exit. Oct 28 05:19:16.224097 systemd-logind[1594]: Removed session 11. Oct 28 05:19:16.258583 containerd[1617]: time="2025-10-28T05:19:16.258355817Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 28 05:19:16.258583 containerd[1617]: time="2025-10-28T05:19:16.258455665Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 28 05:19:16.258805 kubelet[2750]: E1028 05:19:16.258771 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 28 05:19:16.259336 kubelet[2750]: E1028 05:19:16.258825 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 28 05:19:16.259336 kubelet[2750]: E1028 05:19:16.258991 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2mpfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-64d5cb6b45-29rc8_calico-system(fc255c62-41f7-4c05-87d5-240f00279ed4): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:16.260309 kubelet[2750]: E1028 05:19:16.260241 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" podUID="fc255c62-41f7-4c05-87d5-240f00279ed4" Oct 28 05:19:16.694507 containerd[1617]: time="2025-10-28T05:19:16.694187561Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 28 05:19:17.031690 containerd[1617]: time="2025-10-28T05:19:17.031154754Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:17.053126 containerd[1617]: time="2025-10-28T05:19:17.053072984Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 28 05:19:17.053244 containerd[1617]: time="2025-10-28T05:19:17.053155640Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:17.053345 kubelet[2750]: E1028 05:19:17.053300 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 28 05:19:17.053345 kubelet[2750]: E1028 05:19:17.053353 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 28 05:19:17.053554 kubelet[2750]: E1028 05:19:17.053507 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mq9j2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-78lhj_calico-system(855b8ecd-afaf-480c-b1b2-51b12d85eb28): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:17.054851 kubelet[2750]: E1028 05:19:17.054810 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-78lhj" podUID="855b8ecd-afaf-480c-b1b2-51b12d85eb28" Oct 28 05:19:18.695162 containerd[1617]: time="2025-10-28T05:19:18.695053520Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 28 05:19:19.070114 containerd[1617]: time="2025-10-28T05:19:19.070047580Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:19.071140 containerd[1617]: time="2025-10-28T05:19:19.071108261Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 28 05:19:19.071220 containerd[1617]: time="2025-10-28T05:19:19.071145872Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:19.071480 kubelet[2750]: E1028 05:19:19.071388 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:19.071823 kubelet[2750]: E1028 05:19:19.071500 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:19.071823 kubelet[2750]: E1028 05:19:19.071685 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-77w5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7658475cc6-hxjx7_calico-apiserver(1c878ea8-6d56-4615-a865-dddfefeda8f6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:19.072929 kubelet[2750]: E1028 05:19:19.072881 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" podUID="1c878ea8-6d56-4615-a865-dddfefeda8f6" Oct 28 05:19:19.694125 containerd[1617]: time="2025-10-28T05:19:19.694083056Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 28 05:19:20.102163 containerd[1617]: time="2025-10-28T05:19:20.102106522Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:20.130237 containerd[1617]: time="2025-10-28T05:19:20.130169322Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:20.130237 containerd[1617]: time="2025-10-28T05:19:20.130191223Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 28 05:19:20.130524 kubelet[2750]: E1028 05:19:20.130468 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:20.130870 kubelet[2750]: E1028 05:19:20.130534 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:20.130870 kubelet[2750]: E1028 05:19:20.130681 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kmbbw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7658475cc6-2hfhm_calico-apiserver(f075f92a-8aac-4f64-b8e3-246d2217434f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:20.131870 kubelet[2750]: E1028 05:19:20.131830 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" podUID="f075f92a-8aac-4f64-b8e3-246d2217434f" Oct 28 05:19:20.694072 containerd[1617]: time="2025-10-28T05:19:20.694013648Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 28 05:19:21.048714 containerd[1617]: time="2025-10-28T05:19:21.048558442Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:21.049776 containerd[1617]: time="2025-10-28T05:19:21.049718259Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 28 05:19:21.049952 containerd[1617]: time="2025-10-28T05:19:21.049767542Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 28 05:19:21.050011 kubelet[2750]: E1028 05:19:21.049958 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 28 05:19:21.050063 kubelet[2750]: E1028 05:19:21.050025 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 28 05:19:21.050200 kubelet[2750]: E1028 05:19:21.050157 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-chrf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-tfkvb_calico-system(144455de-c836-43b8-85ee-1c567bc354f0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:21.052122 containerd[1617]: time="2025-10-28T05:19:21.052082477Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 28 05:19:21.230657 systemd[1]: Started sshd@11-10.0.0.103:22-10.0.0.1:48376.service - OpenSSH per-connection server daemon (10.0.0.1:48376). Oct 28 05:19:21.305479 sshd[5124]: Accepted publickey for core from 10.0.0.1 port 48376 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:21.307107 sshd-session[5124]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:21.311628 systemd-logind[1594]: New session 12 of user core. Oct 28 05:19:21.319747 systemd[1]: Started session-12.scope - Session 12 of User core. Oct 28 05:19:21.385513 containerd[1617]: time="2025-10-28T05:19:21.385446557Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:21.387550 containerd[1617]: time="2025-10-28T05:19:21.387446608Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 28 05:19:21.387633 containerd[1617]: time="2025-10-28T05:19:21.387562817Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 28 05:19:21.387851 kubelet[2750]: E1028 05:19:21.387797 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 28 05:19:21.388537 kubelet[2750]: E1028 05:19:21.387855 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 28 05:19:21.388537 kubelet[2750]: E1028 05:19:21.387983 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-chrf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-tfkvb_calico-system(144455de-c836-43b8-85ee-1c567bc354f0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:21.389246 kubelet[2750]: E1028 05:19:21.389208 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:19:21.425188 sshd[5127]: Connection closed by 10.0.0.1 port 48376 Oct 28 05:19:21.425543 sshd-session[5124]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:21.438015 systemd[1]: sshd@11-10.0.0.103:22-10.0.0.1:48376.service: Deactivated successfully. Oct 28 05:19:21.441791 systemd[1]: session-12.scope: Deactivated successfully. Oct 28 05:19:21.442602 systemd-logind[1594]: Session 12 logged out. Waiting for processes to exit. Oct 28 05:19:21.446043 systemd[1]: Started sshd@12-10.0.0.103:22-10.0.0.1:48380.service - OpenSSH per-connection server daemon (10.0.0.1:48380). Oct 28 05:19:21.446959 systemd-logind[1594]: Removed session 12. Oct 28 05:19:21.515875 sshd[5141]: Accepted publickey for core from 10.0.0.1 port 48380 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:21.518183 sshd-session[5141]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:21.524013 systemd-logind[1594]: New session 13 of user core. Oct 28 05:19:21.541600 systemd[1]: Started session-13.scope - Session 13 of User core. Oct 28 05:19:21.660922 sshd[5144]: Connection closed by 10.0.0.1 port 48380 Oct 28 05:19:21.664192 sshd-session[5141]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:21.673837 systemd[1]: sshd@12-10.0.0.103:22-10.0.0.1:48380.service: Deactivated successfully. Oct 28 05:19:21.677350 systemd[1]: session-13.scope: Deactivated successfully. Oct 28 05:19:21.679215 systemd-logind[1594]: Session 13 logged out. Waiting for processes to exit. Oct 28 05:19:21.681991 systemd-logind[1594]: Removed session 13. Oct 28 05:19:21.683726 systemd[1]: Started sshd@13-10.0.0.103:22-10.0.0.1:48384.service - OpenSSH per-connection server daemon (10.0.0.1:48384). Oct 28 05:19:21.735880 sshd[5155]: Accepted publickey for core from 10.0.0.1 port 48384 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:21.737311 sshd-session[5155]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:21.741980 systemd-logind[1594]: New session 14 of user core. Oct 28 05:19:21.757622 systemd[1]: Started session-14.scope - Session 14 of User core. Oct 28 05:19:21.836854 sshd[5158]: Connection closed by 10.0.0.1 port 48384 Oct 28 05:19:21.837204 sshd-session[5155]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:21.842115 systemd[1]: sshd@13-10.0.0.103:22-10.0.0.1:48384.service: Deactivated successfully. Oct 28 05:19:21.844645 systemd[1]: session-14.scope: Deactivated successfully. Oct 28 05:19:21.848585 systemd-logind[1594]: Session 14 logged out. Waiting for processes to exit. Oct 28 05:19:21.850146 systemd-logind[1594]: Removed session 14. Oct 28 05:19:26.694922 kubelet[2750]: E1028 05:19:26.694807 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-57bb759ddb-zpxmt" podUID="c69e2212-1789-4672-9606-362def9455d7" Oct 28 05:19:26.859932 systemd[1]: Started sshd@14-10.0.0.103:22-10.0.0.1:48394.service - OpenSSH per-connection server daemon (10.0.0.1:48394). Oct 28 05:19:26.913907 sshd[5176]: Accepted publickey for core from 10.0.0.1 port 48394 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:26.915783 sshd-session[5176]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:26.920317 systemd-logind[1594]: New session 15 of user core. Oct 28 05:19:26.928581 systemd[1]: Started session-15.scope - Session 15 of User core. Oct 28 05:19:26.999870 sshd[5179]: Connection closed by 10.0.0.1 port 48394 Oct 28 05:19:27.000108 sshd-session[5176]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:27.005510 systemd[1]: sshd@14-10.0.0.103:22-10.0.0.1:48394.service: Deactivated successfully. Oct 28 05:19:27.007729 systemd[1]: session-15.scope: Deactivated successfully. Oct 28 05:19:27.008603 systemd-logind[1594]: Session 15 logged out. Waiting for processes to exit. Oct 28 05:19:27.010120 systemd-logind[1594]: Removed session 15. Oct 28 05:19:28.918416 containerd[1617]: time="2025-10-28T05:19:28.918340051Z" level=info msg="TaskExit event in podsandbox handler container_id:\"538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef\" id:\"437ce2c2ac8ae9d4e39bb471387dd3e23acfa9d03c134d0ea926453efed27405\" pid:5213 exited_at:{seconds:1761628768 nanos:917964544}" Oct 28 05:19:28.920510 kubelet[2750]: E1028 05:19:28.920375 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:30.694563 kubelet[2750]: E1028 05:19:30.694505 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" podUID="fc255c62-41f7-4c05-87d5-240f00279ed4" Oct 28 05:19:32.014431 systemd[1]: Started sshd@15-10.0.0.103:22-10.0.0.1:50984.service - OpenSSH per-connection server daemon (10.0.0.1:50984). Oct 28 05:19:32.164230 sshd[5231]: Accepted publickey for core from 10.0.0.1 port 50984 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:32.166034 sshd-session[5231]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:32.170980 systemd-logind[1594]: New session 16 of user core. Oct 28 05:19:32.180577 systemd[1]: Started session-16.scope - Session 16 of User core. Oct 28 05:19:32.410283 sshd[5234]: Connection closed by 10.0.0.1 port 50984 Oct 28 05:19:32.410664 sshd-session[5231]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:32.415002 systemd[1]: sshd@15-10.0.0.103:22-10.0.0.1:50984.service: Deactivated successfully. Oct 28 05:19:32.417141 systemd[1]: session-16.scope: Deactivated successfully. Oct 28 05:19:32.417920 systemd-logind[1594]: Session 16 logged out. Waiting for processes to exit. Oct 28 05:19:32.418999 systemd-logind[1594]: Removed session 16. Oct 28 05:19:32.695069 kubelet[2750]: E1028 05:19:32.694617 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-78lhj" podUID="855b8ecd-afaf-480c-b1b2-51b12d85eb28" Oct 28 05:19:33.693720 kubelet[2750]: E1028 05:19:33.693657 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" podUID="1c878ea8-6d56-4615-a865-dddfefeda8f6" Oct 28 05:19:33.693720 kubelet[2750]: E1028 05:19:33.693678 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" podUID="f075f92a-8aac-4f64-b8e3-246d2217434f" Oct 28 05:19:33.694261 kubelet[2750]: E1028 05:19:33.694007 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:19:37.421749 systemd[1]: Started sshd@16-10.0.0.103:22-10.0.0.1:50992.service - OpenSSH per-connection server daemon (10.0.0.1:50992). Oct 28 05:19:37.476098 sshd[5250]: Accepted publickey for core from 10.0.0.1 port 50992 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:37.477864 sshd-session[5250]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:37.484324 systemd-logind[1594]: New session 17 of user core. Oct 28 05:19:37.489709 systemd[1]: Started session-17.scope - Session 17 of User core. Oct 28 05:19:37.566051 sshd[5253]: Connection closed by 10.0.0.1 port 50992 Oct 28 05:19:37.566356 sshd-session[5250]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:37.571582 systemd[1]: sshd@16-10.0.0.103:22-10.0.0.1:50992.service: Deactivated successfully. Oct 28 05:19:37.574098 systemd[1]: session-17.scope: Deactivated successfully. Oct 28 05:19:37.575010 systemd-logind[1594]: Session 17 logged out. Waiting for processes to exit. Oct 28 05:19:37.576182 systemd-logind[1594]: Removed session 17. Oct 28 05:19:40.694957 containerd[1617]: time="2025-10-28T05:19:40.694771514Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 28 05:19:41.186987 containerd[1617]: time="2025-10-28T05:19:41.186861328Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:41.188705 containerd[1617]: time="2025-10-28T05:19:41.188625437Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 28 05:19:41.188824 containerd[1617]: time="2025-10-28T05:19:41.188667506Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 28 05:19:41.189043 kubelet[2750]: E1028 05:19:41.188984 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 28 05:19:41.189456 kubelet[2750]: E1028 05:19:41.189049 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 28 05:19:41.189456 kubelet[2750]: E1028 05:19:41.189177 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:2f3301ac9dce4d0092d2b57652ea2b6c,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7hr24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-57bb759ddb-zpxmt_calico-system(c69e2212-1789-4672-9606-362def9455d7): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:41.192137 containerd[1617]: time="2025-10-28T05:19:41.192112048Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 28 05:19:41.545676 containerd[1617]: time="2025-10-28T05:19:41.545507576Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:41.547950 containerd[1617]: time="2025-10-28T05:19:41.547904817Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 28 05:19:41.548025 containerd[1617]: time="2025-10-28T05:19:41.547998543Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 28 05:19:41.548251 kubelet[2750]: E1028 05:19:41.548189 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 28 05:19:41.548351 kubelet[2750]: E1028 05:19:41.548265 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 28 05:19:41.548474 kubelet[2750]: E1028 05:19:41.548424 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7hr24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-57bb759ddb-zpxmt_calico-system(c69e2212-1789-4672-9606-362def9455d7): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:41.549665 kubelet[2750]: E1028 05:19:41.549613 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-57bb759ddb-zpxmt" podUID="c69e2212-1789-4672-9606-362def9455d7" Oct 28 05:19:41.694413 containerd[1617]: time="2025-10-28T05:19:41.694338305Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 28 05:19:42.004991 containerd[1617]: time="2025-10-28T05:19:42.004937078Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:42.100823 containerd[1617]: time="2025-10-28T05:19:42.100749083Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 28 05:19:42.100896 containerd[1617]: time="2025-10-28T05:19:42.100813865Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 28 05:19:42.101168 kubelet[2750]: E1028 05:19:42.101108 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 28 05:19:42.101291 kubelet[2750]: E1028 05:19:42.101187 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 28 05:19:42.101542 kubelet[2750]: E1028 05:19:42.101457 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2mpfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-64d5cb6b45-29rc8_calico-system(fc255c62-41f7-4c05-87d5-240f00279ed4): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:42.102725 kubelet[2750]: E1028 05:19:42.102690 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" podUID="fc255c62-41f7-4c05-87d5-240f00279ed4" Oct 28 05:19:42.578244 systemd[1]: Started sshd@17-10.0.0.103:22-10.0.0.1:32894.service - OpenSSH per-connection server daemon (10.0.0.1:32894). Oct 28 05:19:42.645643 sshd[5266]: Accepted publickey for core from 10.0.0.1 port 32894 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:42.647110 sshd-session[5266]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:42.651416 systemd-logind[1594]: New session 18 of user core. Oct 28 05:19:42.660548 systemd[1]: Started session-18.scope - Session 18 of User core. Oct 28 05:19:42.758910 sshd[5269]: Connection closed by 10.0.0.1 port 32894 Oct 28 05:19:42.759853 sshd-session[5266]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:42.771077 systemd[1]: sshd@17-10.0.0.103:22-10.0.0.1:32894.service: Deactivated successfully. Oct 28 05:19:42.773901 systemd[1]: session-18.scope: Deactivated successfully. Oct 28 05:19:42.774995 systemd-logind[1594]: Session 18 logged out. Waiting for processes to exit. Oct 28 05:19:42.780463 systemd[1]: Started sshd@18-10.0.0.103:22-10.0.0.1:32902.service - OpenSSH per-connection server daemon (10.0.0.1:32902). Oct 28 05:19:42.782309 systemd-logind[1594]: Removed session 18. Oct 28 05:19:42.838893 sshd[5284]: Accepted publickey for core from 10.0.0.1 port 32902 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:42.840722 sshd-session[5284]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:42.845756 systemd-logind[1594]: New session 19 of user core. Oct 28 05:19:42.864555 systemd[1]: Started session-19.scope - Session 19 of User core. Oct 28 05:19:43.147082 sshd[5287]: Connection closed by 10.0.0.1 port 32902 Oct 28 05:19:43.147603 sshd-session[5284]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:43.157611 systemd[1]: sshd@18-10.0.0.103:22-10.0.0.1:32902.service: Deactivated successfully. Oct 28 05:19:43.159886 systemd[1]: session-19.scope: Deactivated successfully. Oct 28 05:19:43.160801 systemd-logind[1594]: Session 19 logged out. Waiting for processes to exit. Oct 28 05:19:43.163874 systemd[1]: Started sshd@19-10.0.0.103:22-10.0.0.1:32912.service - OpenSSH per-connection server daemon (10.0.0.1:32912). Oct 28 05:19:43.165156 systemd-logind[1594]: Removed session 19. Oct 28 05:19:43.227022 sshd[5299]: Accepted publickey for core from 10.0.0.1 port 32912 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:43.229102 sshd-session[5299]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:43.234448 systemd-logind[1594]: New session 20 of user core. Oct 28 05:19:43.247643 systemd[1]: Started session-20.scope - Session 20 of User core. Oct 28 05:19:43.888543 sshd[5302]: Connection closed by 10.0.0.1 port 32912 Oct 28 05:19:43.889094 sshd-session[5299]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:43.905937 systemd[1]: sshd@19-10.0.0.103:22-10.0.0.1:32912.service: Deactivated successfully. Oct 28 05:19:43.910827 systemd[1]: session-20.scope: Deactivated successfully. Oct 28 05:19:43.911973 systemd-logind[1594]: Session 20 logged out. Waiting for processes to exit. Oct 28 05:19:43.916498 systemd[1]: Started sshd@20-10.0.0.103:22-10.0.0.1:32916.service - OpenSSH per-connection server daemon (10.0.0.1:32916). Oct 28 05:19:43.917453 systemd-logind[1594]: Removed session 20. Oct 28 05:19:43.964282 sshd[5321]: Accepted publickey for core from 10.0.0.1 port 32916 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:43.966640 sshd-session[5321]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:43.971554 systemd-logind[1594]: New session 21 of user core. Oct 28 05:19:43.982712 systemd[1]: Started session-21.scope - Session 21 of User core. Oct 28 05:19:44.179202 sshd[5324]: Connection closed by 10.0.0.1 port 32916 Oct 28 05:19:44.179577 sshd-session[5321]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:44.190487 systemd[1]: sshd@20-10.0.0.103:22-10.0.0.1:32916.service: Deactivated successfully. Oct 28 05:19:44.193178 systemd[1]: session-21.scope: Deactivated successfully. Oct 28 05:19:44.194039 systemd-logind[1594]: Session 21 logged out. Waiting for processes to exit. Oct 28 05:19:44.197267 systemd[1]: Started sshd@21-10.0.0.103:22-10.0.0.1:32918.service - OpenSSH per-connection server daemon (10.0.0.1:32918). Oct 28 05:19:44.198649 systemd-logind[1594]: Removed session 21. Oct 28 05:19:44.243185 sshd[5336]: Accepted publickey for core from 10.0.0.1 port 32918 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:44.244510 sshd-session[5336]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:44.249081 systemd-logind[1594]: New session 22 of user core. Oct 28 05:19:44.258542 systemd[1]: Started session-22.scope - Session 22 of User core. Oct 28 05:19:44.336897 sshd[5339]: Connection closed by 10.0.0.1 port 32918 Oct 28 05:19:44.337216 sshd-session[5336]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:44.341541 systemd[1]: sshd@21-10.0.0.103:22-10.0.0.1:32918.service: Deactivated successfully. Oct 28 05:19:44.343698 systemd[1]: session-22.scope: Deactivated successfully. Oct 28 05:19:44.344612 systemd-logind[1594]: Session 22 logged out. Waiting for processes to exit. Oct 28 05:19:44.346191 systemd-logind[1594]: Removed session 22. Oct 28 05:19:45.695137 containerd[1617]: time="2025-10-28T05:19:45.694976419Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 28 05:19:46.087690 containerd[1617]: time="2025-10-28T05:19:46.087616451Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:46.107210 containerd[1617]: time="2025-10-28T05:19:46.107038438Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 28 05:19:46.107498 kubelet[2750]: E1028 05:19:46.107420 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:46.107916 kubelet[2750]: E1028 05:19:46.107502 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:46.107916 kubelet[2750]: E1028 05:19:46.107646 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kmbbw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7658475cc6-2hfhm_calico-apiserver(f075f92a-8aac-4f64-b8e3-246d2217434f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:46.109240 kubelet[2750]: E1028 05:19:46.109204 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" podUID="f075f92a-8aac-4f64-b8e3-246d2217434f" Oct 28 05:19:46.133705 containerd[1617]: time="2025-10-28T05:19:46.106881683Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:46.695221 containerd[1617]: time="2025-10-28T05:19:46.694920810Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 28 05:19:47.045258 containerd[1617]: time="2025-10-28T05:19:47.045122688Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:47.054995 containerd[1617]: time="2025-10-28T05:19:47.054941348Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 28 05:19:47.055103 containerd[1617]: time="2025-10-28T05:19:47.055026569Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:47.055273 kubelet[2750]: E1028 05:19:47.055177 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 28 05:19:47.055273 kubelet[2750]: E1028 05:19:47.055230 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 28 05:19:47.055466 kubelet[2750]: E1028 05:19:47.055362 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mq9j2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-78lhj_calico-system(855b8ecd-afaf-480c-b1b2-51b12d85eb28): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:47.056636 kubelet[2750]: E1028 05:19:47.056564 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-78lhj" podUID="855b8ecd-afaf-480c-b1b2-51b12d85eb28" Oct 28 05:19:47.694049 kubelet[2750]: E1028 05:19:47.693263 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:47.695088 containerd[1617]: time="2025-10-28T05:19:47.694000532Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 28 05:19:48.089688 containerd[1617]: time="2025-10-28T05:19:48.089511025Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:48.090978 containerd[1617]: time="2025-10-28T05:19:48.090931798Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 28 05:19:48.091107 containerd[1617]: time="2025-10-28T05:19:48.091037245Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 28 05:19:48.091278 kubelet[2750]: E1028 05:19:48.091209 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:48.091278 kubelet[2750]: E1028 05:19:48.091274 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 28 05:19:48.091611 kubelet[2750]: E1028 05:19:48.091524 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-77w5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7658475cc6-hxjx7_calico-apiserver(1c878ea8-6d56-4615-a865-dddfefeda8f6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:48.092083 containerd[1617]: time="2025-10-28T05:19:48.091948871Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 28 05:19:48.093195 kubelet[2750]: E1028 05:19:48.093142 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" podUID="1c878ea8-6d56-4615-a865-dddfefeda8f6" Oct 28 05:19:48.400183 containerd[1617]: time="2025-10-28T05:19:48.400029499Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:48.448226 containerd[1617]: time="2025-10-28T05:19:48.448159409Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 28 05:19:48.448226 containerd[1617]: time="2025-10-28T05:19:48.448206748Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 28 05:19:48.448527 kubelet[2750]: E1028 05:19:48.448475 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 28 05:19:48.448582 kubelet[2750]: E1028 05:19:48.448543 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 28 05:19:48.448709 kubelet[2750]: E1028 05:19:48.448669 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-chrf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-tfkvb_calico-system(144455de-c836-43b8-85ee-1c567bc354f0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:48.451523 containerd[1617]: time="2025-10-28T05:19:48.451460327Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 28 05:19:48.824710 containerd[1617]: time="2025-10-28T05:19:48.824661375Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 28 05:19:48.883459 containerd[1617]: time="2025-10-28T05:19:48.883386133Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 28 05:19:48.883459 containerd[1617]: time="2025-10-28T05:19:48.883434905Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 28 05:19:48.883779 kubelet[2750]: E1028 05:19:48.883727 2750 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 28 05:19:48.884177 kubelet[2750]: E1028 05:19:48.883795 2750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 28 05:19:48.884177 kubelet[2750]: E1028 05:19:48.883949 2750 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-chrf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-tfkvb_calico-system(144455de-c836-43b8-85ee-1c567bc354f0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 28 05:19:48.885952 kubelet[2750]: E1028 05:19:48.885916 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:19:49.349989 systemd[1]: Started sshd@22-10.0.0.103:22-10.0.0.1:32934.service - OpenSSH per-connection server daemon (10.0.0.1:32934). Oct 28 05:19:49.406997 sshd[5360]: Accepted publickey for core from 10.0.0.1 port 32934 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:49.408741 sshd-session[5360]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:49.416115 systemd-logind[1594]: New session 23 of user core. Oct 28 05:19:49.420566 systemd[1]: Started session-23.scope - Session 23 of User core. Oct 28 05:19:49.498289 sshd[5363]: Connection closed by 10.0.0.1 port 32934 Oct 28 05:19:49.498654 sshd-session[5360]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:49.503481 systemd[1]: sshd@22-10.0.0.103:22-10.0.0.1:32934.service: Deactivated successfully. Oct 28 05:19:49.505626 systemd[1]: session-23.scope: Deactivated successfully. Oct 28 05:19:49.506593 systemd-logind[1594]: Session 23 logged out. Waiting for processes to exit. Oct 28 05:19:49.508807 systemd-logind[1594]: Removed session 23. Oct 28 05:19:54.510263 systemd[1]: Started sshd@23-10.0.0.103:22-10.0.0.1:39442.service - OpenSSH per-connection server daemon (10.0.0.1:39442). Oct 28 05:19:54.555789 sshd[5378]: Accepted publickey for core from 10.0.0.1 port 39442 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:54.557267 sshd-session[5378]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:54.562036 systemd-logind[1594]: New session 24 of user core. Oct 28 05:19:54.576565 systemd[1]: Started session-24.scope - Session 24 of User core. Oct 28 05:19:54.647537 sshd[5381]: Connection closed by 10.0.0.1 port 39442 Oct 28 05:19:54.647851 sshd-session[5378]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:54.652121 systemd[1]: sshd@23-10.0.0.103:22-10.0.0.1:39442.service: Deactivated successfully. Oct 28 05:19:54.654970 systemd[1]: session-24.scope: Deactivated successfully. Oct 28 05:19:54.655898 systemd-logind[1594]: Session 24 logged out. Waiting for processes to exit. Oct 28 05:19:54.657390 systemd-logind[1594]: Removed session 24. Oct 28 05:19:54.693105 kubelet[2750]: E1028 05:19:54.693076 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:54.693448 kubelet[2750]: E1028 05:19:54.693220 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 28 05:19:56.694711 kubelet[2750]: E1028 05:19:56.694598 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-2hfhm" podUID="f075f92a-8aac-4f64-b8e3-246d2217434f" Oct 28 05:19:56.694711 kubelet[2750]: E1028 05:19:56.694633 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-64d5cb6b45-29rc8" podUID="fc255c62-41f7-4c05-87d5-240f00279ed4" Oct 28 05:19:56.695980 kubelet[2750]: E1028 05:19:56.695906 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-57bb759ddb-zpxmt" podUID="c69e2212-1789-4672-9606-362def9455d7" Oct 28 05:19:58.694748 kubelet[2750]: E1028 05:19:58.694554 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-78lhj" podUID="855b8ecd-afaf-480c-b1b2-51b12d85eb28" Oct 28 05:19:58.938853 containerd[1617]: time="2025-10-28T05:19:58.938801010Z" level=info msg="TaskExit event in podsandbox handler container_id:\"538913a668d09884080199a3b2c51da8180061769c8f68b6e30de1e45a43c0ef\" id:\"83778871e4debff7dd1314038c8d3101765ab1b624cbf0f77b4c0b219df8004d\" pid:5404 exited_at:{seconds:1761628798 nanos:938413192}" Oct 28 05:19:59.660778 systemd[1]: Started sshd@24-10.0.0.103:22-10.0.0.1:39446.service - OpenSSH per-connection server daemon (10.0.0.1:39446). Oct 28 05:19:59.747256 sshd[5419]: Accepted publickey for core from 10.0.0.1 port 39446 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:19:59.749371 sshd-session[5419]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:19:59.754088 systemd-logind[1594]: New session 25 of user core. Oct 28 05:19:59.764538 systemd[1]: Started session-25.scope - Session 25 of User core. Oct 28 05:19:59.847663 sshd[5422]: Connection closed by 10.0.0.1 port 39446 Oct 28 05:19:59.847988 sshd-session[5419]: pam_unix(sshd:session): session closed for user core Oct 28 05:19:59.854067 systemd[1]: sshd@24-10.0.0.103:22-10.0.0.1:39446.service: Deactivated successfully. Oct 28 05:19:59.857205 systemd[1]: session-25.scope: Deactivated successfully. Oct 28 05:19:59.858445 systemd-logind[1594]: Session 25 logged out. Waiting for processes to exit. Oct 28 05:19:59.860267 systemd-logind[1594]: Removed session 25. Oct 28 05:20:00.696274 kubelet[2750]: E1028 05:20:00.696209 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-tfkvb" podUID="144455de-c836-43b8-85ee-1c567bc354f0" Oct 28 05:20:02.693431 kubelet[2750]: E1028 05:20:02.693351 2750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7658475cc6-hxjx7" podUID="1c878ea8-6d56-4615-a865-dddfefeda8f6" Oct 28 05:20:04.862139 systemd[1]: Started sshd@25-10.0.0.103:22-10.0.0.1:60260.service - OpenSSH per-connection server daemon (10.0.0.1:60260). Oct 28 05:20:04.933748 sshd[5435]: Accepted publickey for core from 10.0.0.1 port 60260 ssh2: RSA SHA256:fnPK7ZFusTyP70xk0T3ldjOg8cdR8K1Md9pjQBlXC4c Oct 28 05:20:04.935463 sshd-session[5435]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 28 05:20:04.939963 systemd-logind[1594]: New session 26 of user core. Oct 28 05:20:04.946529 systemd[1]: Started session-26.scope - Session 26 of User core. Oct 28 05:20:05.029260 sshd[5438]: Connection closed by 10.0.0.1 port 60260 Oct 28 05:20:05.029706 sshd-session[5435]: pam_unix(sshd:session): session closed for user core Oct 28 05:20:05.034493 systemd-logind[1594]: Session 26 logged out. Waiting for processes to exit. Oct 28 05:20:05.034911 systemd[1]: sshd@25-10.0.0.103:22-10.0.0.1:60260.service: Deactivated successfully. Oct 28 05:20:05.037314 systemd[1]: session-26.scope: Deactivated successfully. Oct 28 05:20:05.039509 systemd-logind[1594]: Removed session 26. Oct 28 05:20:05.693642 kubelet[2750]: E1028 05:20:05.693593 2750 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"