Nov 24 00:27:00.958750 kernel: Linux version 6.12.58-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Sun Nov 23 20:54:38 -00 2025 Nov 24 00:27:00.958784 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=1969a6ee0c0ec5507eb68849c160e94c58e52d2291c767873af68a1f52b30801 Nov 24 00:27:00.958797 kernel: BIOS-provided physical RAM map: Nov 24 00:27:00.958807 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009f7ff] usable Nov 24 00:27:00.958815 kernel: BIOS-e820: [mem 0x000000000009f800-0x000000000009ffff] reserved Nov 24 00:27:00.958825 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Nov 24 00:27:00.958838 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdcfff] usable Nov 24 00:27:00.958848 kernel: BIOS-e820: [mem 0x000000007ffdd000-0x000000007fffffff] reserved Nov 24 00:27:00.958858 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Nov 24 00:27:00.958864 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Nov 24 00:27:00.958870 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Nov 24 00:27:00.958876 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Nov 24 00:27:00.958882 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000017fffffff] usable Nov 24 00:27:00.958889 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Nov 24 00:27:00.958898 kernel: NX (Execute Disable) protection: active Nov 24 00:27:00.958905 kernel: APIC: Static calls initialized Nov 24 00:27:00.958911 kernel: SMBIOS 2.8 present. Nov 24 00:27:00.958918 kernel: DMI: Linode Compute Instance/Standard PC (Q35 + ICH9, 2009), BIOS Not Specified Nov 24 00:27:00.958925 kernel: DMI: Memory slots populated: 1/1 Nov 24 00:27:00.958931 kernel: Hypervisor detected: KVM Nov 24 00:27:00.958940 kernel: last_pfn = 0x7ffdd max_arch_pfn = 0x400000000 Nov 24 00:27:00.958946 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Nov 24 00:27:00.958953 kernel: kvm-clock: using sched offset of 7189961185 cycles Nov 24 00:27:00.958960 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Nov 24 00:27:00.958967 kernel: tsc: Detected 1999.999 MHz processor Nov 24 00:27:00.958974 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Nov 24 00:27:00.958981 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Nov 24 00:27:00.958988 kernel: last_pfn = 0x180000 max_arch_pfn = 0x400000000 Nov 24 00:27:00.958995 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Nov 24 00:27:00.959002 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Nov 24 00:27:00.959011 kernel: last_pfn = 0x7ffdd max_arch_pfn = 0x400000000 Nov 24 00:27:00.959017 kernel: Using GB pages for direct mapping Nov 24 00:27:00.959024 kernel: ACPI: Early table checksum verification disabled Nov 24 00:27:00.959031 kernel: ACPI: RSDP 0x00000000000F5160 000014 (v00 BOCHS ) Nov 24 00:27:00.959037 kernel: ACPI: RSDT 0x000000007FFE2307 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:27:00.959044 kernel: ACPI: FACP 0x000000007FFE20F7 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:27:00.959051 kernel: ACPI: DSDT 0x000000007FFE0040 0020B7 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:27:00.959058 kernel: ACPI: FACS 0x000000007FFE0000 000040 Nov 24 00:27:00.959064 kernel: ACPI: APIC 0x000000007FFE21EB 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:27:00.959074 kernel: ACPI: HPET 0x000000007FFE226B 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:27:00.959083 kernel: ACPI: MCFG 0x000000007FFE22A3 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:27:00.959090 kernel: ACPI: WAET 0x000000007FFE22DF 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:27:00.959097 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe20f7-0x7ffe21ea] Nov 24 00:27:00.959104 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe20f6] Nov 24 00:27:00.959112 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Nov 24 00:27:00.959119 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe21eb-0x7ffe226a] Nov 24 00:27:00.959126 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe226b-0x7ffe22a2] Nov 24 00:27:00.959133 kernel: ACPI: Reserving MCFG table memory at [mem 0x7ffe22a3-0x7ffe22de] Nov 24 00:27:00.959139 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe22df-0x7ffe2306] Nov 24 00:27:00.959146 kernel: No NUMA configuration found Nov 24 00:27:00.959152 kernel: Faking a node at [mem 0x0000000000000000-0x000000017fffffff] Nov 24 00:27:00.959159 kernel: NODE_DATA(0) allocated [mem 0x17fff6dc0-0x17fffdfff] Nov 24 00:27:00.959166 kernel: Zone ranges: Nov 24 00:27:00.959175 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Nov 24 00:27:00.959181 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Nov 24 00:27:00.959188 kernel: Normal [mem 0x0000000100000000-0x000000017fffffff] Nov 24 00:27:00.959195 kernel: Device empty Nov 24 00:27:00.959201 kernel: Movable zone start for each node Nov 24 00:27:00.959208 kernel: Early memory node ranges Nov 24 00:27:00.959215 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Nov 24 00:27:00.959221 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdcfff] Nov 24 00:27:00.959228 kernel: node 0: [mem 0x0000000100000000-0x000000017fffffff] Nov 24 00:27:00.959235 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000017fffffff] Nov 24 00:27:00.959243 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Nov 24 00:27:00.959250 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Nov 24 00:27:00.959257 kernel: On node 0, zone Normal: 35 pages in unavailable ranges Nov 24 00:27:00.959264 kernel: ACPI: PM-Timer IO Port: 0x608 Nov 24 00:27:00.959270 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Nov 24 00:27:00.959277 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Nov 24 00:27:00.959284 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Nov 24 00:27:00.959290 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Nov 24 00:27:00.959297 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Nov 24 00:27:00.959306 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Nov 24 00:27:00.959312 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Nov 24 00:27:00.959319 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Nov 24 00:27:00.959326 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Nov 24 00:27:00.959332 kernel: TSC deadline timer available Nov 24 00:27:00.959339 kernel: CPU topo: Max. logical packages: 1 Nov 24 00:27:00.959346 kernel: CPU topo: Max. logical dies: 1 Nov 24 00:27:00.959352 kernel: CPU topo: Max. dies per package: 1 Nov 24 00:27:00.959359 kernel: CPU topo: Max. threads per core: 1 Nov 24 00:27:00.959368 kernel: CPU topo: Num. cores per package: 2 Nov 24 00:27:00.959374 kernel: CPU topo: Num. threads per package: 2 Nov 24 00:27:00.959381 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Nov 24 00:27:00.959388 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Nov 24 00:27:00.959394 kernel: kvm-guest: KVM setup pv remote TLB flush Nov 24 00:27:00.959401 kernel: kvm-guest: setup PV sched yield Nov 24 00:27:00.959407 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Nov 24 00:27:00.959414 kernel: Booting paravirtualized kernel on KVM Nov 24 00:27:00.959421 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Nov 24 00:27:00.959430 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Nov 24 00:27:00.959437 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Nov 24 00:27:00.959443 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Nov 24 00:27:00.959450 kernel: pcpu-alloc: [0] 0 1 Nov 24 00:27:00.959456 kernel: kvm-guest: PV spinlocks enabled Nov 24 00:27:00.959463 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Nov 24 00:27:00.959471 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=1969a6ee0c0ec5507eb68849c160e94c58e52d2291c767873af68a1f52b30801 Nov 24 00:27:00.959478 kernel: random: crng init done Nov 24 00:27:00.959487 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Nov 24 00:27:00.959493 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 24 00:27:00.959500 kernel: Fallback order for Node 0: 0 Nov 24 00:27:00.959507 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1048443 Nov 24 00:27:00.959513 kernel: Policy zone: Normal Nov 24 00:27:00.959520 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Nov 24 00:27:00.959527 kernel: software IO TLB: area num 2. Nov 24 00:27:00.959533 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Nov 24 00:27:00.959540 kernel: ftrace: allocating 40103 entries in 157 pages Nov 24 00:27:00.959549 kernel: ftrace: allocated 157 pages with 5 groups Nov 24 00:27:00.959555 kernel: Dynamic Preempt: voluntary Nov 24 00:27:00.959562 kernel: rcu: Preemptible hierarchical RCU implementation. Nov 24 00:27:00.959570 kernel: rcu: RCU event tracing is enabled. Nov 24 00:27:00.959577 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Nov 24 00:27:00.959584 kernel: Trampoline variant of Tasks RCU enabled. Nov 24 00:27:00.959590 kernel: Rude variant of Tasks RCU enabled. Nov 24 00:27:00.959597 kernel: Tracing variant of Tasks RCU enabled. Nov 24 00:27:00.959604 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Nov 24 00:27:00.959611 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Nov 24 00:27:00.959619 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Nov 24 00:27:00.959633 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Nov 24 00:27:00.959643 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Nov 24 00:27:00.959650 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Nov 24 00:27:00.959662 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Nov 24 00:27:00.959674 kernel: Console: colour VGA+ 80x25 Nov 24 00:27:00.959686 kernel: printk: legacy console [tty0] enabled Nov 24 00:27:00.959697 kernel: printk: legacy console [ttyS0] enabled Nov 24 00:27:00.959730 kernel: ACPI: Core revision 20240827 Nov 24 00:27:00.959743 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Nov 24 00:27:00.959752 kernel: APIC: Switch to symmetric I/O mode setup Nov 24 00:27:00.959764 kernel: x2apic enabled Nov 24 00:27:00.959776 kernel: APIC: Switched APIC routing to: physical x2apic Nov 24 00:27:00.959788 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Nov 24 00:27:00.959800 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Nov 24 00:27:00.959808 kernel: kvm-guest: setup PV IPIs Nov 24 00:27:00.959818 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Nov 24 00:27:00.959825 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x39a85afc727, max_idle_ns: 881590685098 ns Nov 24 00:27:00.959832 kernel: Calibrating delay loop (skipped) preset value.. 3999.99 BogoMIPS (lpj=1999999) Nov 24 00:27:00.959839 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Nov 24 00:27:00.959849 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Nov 24 00:27:00.959861 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Nov 24 00:27:00.959873 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Nov 24 00:27:00.959885 kernel: Spectre V2 : Mitigation: Retpolines Nov 24 00:27:00.959896 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Nov 24 00:27:00.959908 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Nov 24 00:27:00.959915 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Nov 24 00:27:00.959923 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Nov 24 00:27:00.959930 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Nov 24 00:27:00.959940 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Nov 24 00:27:00.959951 kernel: active return thunk: srso_alias_return_thunk Nov 24 00:27:00.959963 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Nov 24 00:27:00.959975 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Nov 24 00:27:00.959989 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Nov 24 00:27:00.960000 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Nov 24 00:27:00.960012 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Nov 24 00:27:00.960020 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Nov 24 00:27:00.960028 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Nov 24 00:27:00.960040 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Nov 24 00:27:00.960051 kernel: x86/fpu: xstate_offset[9]: 832, xstate_sizes[9]: 8 Nov 24 00:27:00.960063 kernel: x86/fpu: Enabled xstate features 0x207, context size is 840 bytes, using 'compacted' format. Nov 24 00:27:00.960074 kernel: Freeing SMP alternatives memory: 32K Nov 24 00:27:00.960089 kernel: pid_max: default: 32768 minimum: 301 Nov 24 00:27:00.960100 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Nov 24 00:27:00.960111 kernel: landlock: Up and running. Nov 24 00:27:00.960123 kernel: SELinux: Initializing. Nov 24 00:27:00.960134 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 24 00:27:00.960146 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 24 00:27:00.960158 kernel: smpboot: CPU0: AMD EPYC 7713 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Nov 24 00:27:00.960169 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Nov 24 00:27:00.960180 kernel: ... version: 0 Nov 24 00:27:00.960194 kernel: ... bit width: 48 Nov 24 00:27:00.960206 kernel: ... generic registers: 6 Nov 24 00:27:00.960217 kernel: ... value mask: 0000ffffffffffff Nov 24 00:27:00.960229 kernel: ... max period: 00007fffffffffff Nov 24 00:27:00.960240 kernel: ... fixed-purpose events: 0 Nov 24 00:27:00.960251 kernel: ... event mask: 000000000000003f Nov 24 00:27:00.960262 kernel: signal: max sigframe size: 3376 Nov 24 00:27:00.960273 kernel: rcu: Hierarchical SRCU implementation. Nov 24 00:27:00.960285 kernel: rcu: Max phase no-delay instances is 400. Nov 24 00:27:00.960299 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Nov 24 00:27:00.960310 kernel: smp: Bringing up secondary CPUs ... Nov 24 00:27:00.960318 kernel: smpboot: x86: Booting SMP configuration: Nov 24 00:27:00.960325 kernel: .... node #0, CPUs: #1 Nov 24 00:27:00.960332 kernel: smp: Brought up 1 node, 2 CPUs Nov 24 00:27:00.960339 kernel: smpboot: Total of 2 processors activated (7999.99 BogoMIPS) Nov 24 00:27:00.960347 kernel: Memory: 3952856K/4193772K available (14336K kernel code, 2444K rwdata, 26064K rodata, 46188K init, 2572K bss, 235488K reserved, 0K cma-reserved) Nov 24 00:27:00.960354 kernel: devtmpfs: initialized Nov 24 00:27:00.960361 kernel: x86/mm: Memory block size: 128MB Nov 24 00:27:00.960370 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Nov 24 00:27:00.960378 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Nov 24 00:27:00.960385 kernel: pinctrl core: initialized pinctrl subsystem Nov 24 00:27:00.960392 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Nov 24 00:27:00.960399 kernel: audit: initializing netlink subsys (disabled) Nov 24 00:27:00.960406 kernel: audit: type=2000 audit(1763944017.944:1): state=initialized audit_enabled=0 res=1 Nov 24 00:27:00.960413 kernel: thermal_sys: Registered thermal governor 'step_wise' Nov 24 00:27:00.960420 kernel: thermal_sys: Registered thermal governor 'user_space' Nov 24 00:27:00.960427 kernel: cpuidle: using governor menu Nov 24 00:27:00.960436 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Nov 24 00:27:00.960443 kernel: dca service started, version 1.12.1 Nov 24 00:27:00.960450 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] (base 0xb0000000) for domain 0000 [bus 00-ff] Nov 24 00:27:00.960457 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] reserved as E820 entry Nov 24 00:27:00.960464 kernel: PCI: Using configuration type 1 for base access Nov 24 00:27:00.960471 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Nov 24 00:27:00.960478 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Nov 24 00:27:00.960485 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Nov 24 00:27:00.960492 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Nov 24 00:27:00.960501 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Nov 24 00:27:00.960508 kernel: ACPI: Added _OSI(Module Device) Nov 24 00:27:00.960515 kernel: ACPI: Added _OSI(Processor Device) Nov 24 00:27:00.960522 kernel: ACPI: Added _OSI(Processor Aggregator Device) Nov 24 00:27:00.960528 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Nov 24 00:27:00.960535 kernel: ACPI: Interpreter enabled Nov 24 00:27:00.960542 kernel: ACPI: PM: (supports S0 S3 S5) Nov 24 00:27:00.960549 kernel: ACPI: Using IOAPIC for interrupt routing Nov 24 00:27:00.960557 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Nov 24 00:27:00.960566 kernel: PCI: Using E820 reservations for host bridge windows Nov 24 00:27:00.960573 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Nov 24 00:27:00.960580 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Nov 24 00:27:00.964861 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Nov 24 00:27:00.965064 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Nov 24 00:27:00.965197 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Nov 24 00:27:00.965207 kernel: PCI host bridge to bus 0000:00 Nov 24 00:27:00.965347 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Nov 24 00:27:00.965461 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Nov 24 00:27:00.965636 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Nov 24 00:27:00.965773 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xafffffff window] Nov 24 00:27:00.965925 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Nov 24 00:27:00.966084 kernel: pci_bus 0000:00: root bus resource [mem 0x180000000-0x97fffffff window] Nov 24 00:27:00.966244 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Nov 24 00:27:00.966446 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Nov 24 00:27:00.966638 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Nov 24 00:27:00.971381 kernel: pci 0000:00:01.0: BAR 0 [mem 0xfd000000-0xfdffffff pref] Nov 24 00:27:00.971516 kernel: pci 0000:00:01.0: BAR 2 [mem 0xfebd0000-0xfebd0fff] Nov 24 00:27:00.971652 kernel: pci 0000:00:01.0: ROM [mem 0xfebc0000-0xfebcffff pref] Nov 24 00:27:00.971820 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Nov 24 00:27:00.971964 kernel: pci 0000:00:02.0: [1af4:1004] type 00 class 0x010000 conventional PCI endpoint Nov 24 00:27:00.972095 kernel: pci 0000:00:02.0: BAR 0 [io 0xc000-0xc03f] Nov 24 00:27:00.972216 kernel: pci 0000:00:02.0: BAR 1 [mem 0xfebd1000-0xfebd1fff] Nov 24 00:27:00.972336 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfe000000-0xfe003fff 64bit pref] Nov 24 00:27:00.972466 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Nov 24 00:27:00.972588 kernel: pci 0000:00:03.0: BAR 0 [io 0xc040-0xc07f] Nov 24 00:27:00.972731 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebd2000-0xfebd2fff] Nov 24 00:27:00.972867 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe004000-0xfe007fff 64bit pref] Nov 24 00:27:00.972989 kernel: pci 0000:00:03.0: ROM [mem 0xfeb80000-0xfebbffff pref] Nov 24 00:27:00.973118 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Nov 24 00:27:00.973240 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Nov 24 00:27:00.973372 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Nov 24 00:27:00.973493 kernel: pci 0000:00:1f.2: BAR 4 [io 0xc0c0-0xc0df] Nov 24 00:27:00.973612 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xfebd3000-0xfebd3fff] Nov 24 00:27:00.973801 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Nov 24 00:27:00.973948 kernel: pci 0000:00:1f.3: BAR 4 [io 0x0700-0x073f] Nov 24 00:27:00.973960 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Nov 24 00:27:00.973968 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Nov 24 00:27:00.973976 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Nov 24 00:27:00.973983 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Nov 24 00:27:00.973991 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Nov 24 00:27:00.973998 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Nov 24 00:27:00.974010 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Nov 24 00:27:00.974018 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Nov 24 00:27:00.974025 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Nov 24 00:27:00.974033 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Nov 24 00:27:00.974040 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Nov 24 00:27:00.974047 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Nov 24 00:27:00.974055 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Nov 24 00:27:00.974062 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Nov 24 00:27:00.974070 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Nov 24 00:27:00.974080 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Nov 24 00:27:00.974087 kernel: iommu: Default domain type: Translated Nov 24 00:27:00.974095 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Nov 24 00:27:00.974102 kernel: PCI: Using ACPI for IRQ routing Nov 24 00:27:00.974110 kernel: PCI: pci_cache_line_size set to 64 bytes Nov 24 00:27:00.974117 kernel: e820: reserve RAM buffer [mem 0x0009f800-0x0009ffff] Nov 24 00:27:00.974133 kernel: e820: reserve RAM buffer [mem 0x7ffdd000-0x7fffffff] Nov 24 00:27:00.974264 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Nov 24 00:27:00.974390 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Nov 24 00:27:00.974510 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Nov 24 00:27:00.974520 kernel: vgaarb: loaded Nov 24 00:27:00.974528 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Nov 24 00:27:00.974536 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Nov 24 00:27:00.974543 kernel: clocksource: Switched to clocksource kvm-clock Nov 24 00:27:00.974550 kernel: VFS: Disk quotas dquot_6.6.0 Nov 24 00:27:00.974558 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Nov 24 00:27:00.974566 kernel: pnp: PnP ACPI init Nov 24 00:27:00.974702 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Nov 24 00:27:00.977186 kernel: pnp: PnP ACPI: found 5 devices Nov 24 00:27:00.977207 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Nov 24 00:27:00.977220 kernel: NET: Registered PF_INET protocol family Nov 24 00:27:00.977231 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 24 00:27:00.977242 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Nov 24 00:27:00.977252 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Nov 24 00:27:00.977263 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Nov 24 00:27:00.977291 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Nov 24 00:27:00.977302 kernel: TCP: Hash tables configured (established 32768 bind 32768) Nov 24 00:27:00.977311 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 24 00:27:00.977320 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 24 00:27:00.977330 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Nov 24 00:27:00.977340 kernel: NET: Registered PF_XDP protocol family Nov 24 00:27:00.977631 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Nov 24 00:27:00.977822 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Nov 24 00:27:00.977979 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Nov 24 00:27:00.978127 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xafffffff window] Nov 24 00:27:00.978253 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Nov 24 00:27:00.978364 kernel: pci_bus 0000:00: resource 9 [mem 0x180000000-0x97fffffff window] Nov 24 00:27:00.978375 kernel: PCI: CLS 0 bytes, default 64 Nov 24 00:27:00.978384 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Nov 24 00:27:00.978391 kernel: software IO TLB: mapped [mem 0x000000007bfdd000-0x000000007ffdd000] (64MB) Nov 24 00:27:00.978400 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x39a85afc727, max_idle_ns: 881590685098 ns Nov 24 00:27:00.978409 kernel: Initialise system trusted keyrings Nov 24 00:27:00.978421 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Nov 24 00:27:00.978429 kernel: Key type asymmetric registered Nov 24 00:27:00.978437 kernel: Asymmetric key parser 'x509' registered Nov 24 00:27:00.978445 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Nov 24 00:27:00.978453 kernel: io scheduler mq-deadline registered Nov 24 00:27:00.978461 kernel: io scheduler kyber registered Nov 24 00:27:00.978469 kernel: io scheduler bfq registered Nov 24 00:27:00.978477 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Nov 24 00:27:00.978486 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Nov 24 00:27:00.978498 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Nov 24 00:27:00.978505 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Nov 24 00:27:00.978513 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Nov 24 00:27:00.978521 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Nov 24 00:27:00.978529 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Nov 24 00:27:00.978537 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Nov 24 00:27:00.979097 kernel: rtc_cmos 00:03: RTC can wake from S4 Nov 24 00:27:00.979114 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Nov 24 00:27:00.979241 kernel: rtc_cmos 00:03: registered as rtc0 Nov 24 00:27:00.979547 kernel: rtc_cmos 00:03: setting system clock to 2025-11-24T00:27:00 UTC (1763944020) Nov 24 00:27:00.979662 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Nov 24 00:27:00.979672 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Nov 24 00:27:00.979681 kernel: NET: Registered PF_INET6 protocol family Nov 24 00:27:00.979689 kernel: Segment Routing with IPv6 Nov 24 00:27:00.979696 kernel: In-situ OAM (IOAM) with IPv6 Nov 24 00:27:00.979704 kernel: NET: Registered PF_PACKET protocol family Nov 24 00:27:00.979728 kernel: Key type dns_resolver registered Nov 24 00:27:00.979740 kernel: IPI shorthand broadcast: enabled Nov 24 00:27:00.979748 kernel: sched_clock: Marking stable (2922007287, 349609611)->(3371181705, -99564807) Nov 24 00:27:00.979756 kernel: registered taskstats version 1 Nov 24 00:27:00.979763 kernel: Loading compiled-in X.509 certificates Nov 24 00:27:00.979771 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.58-flatcar: 5d380f93d180914be04be8068ab300f495c35900' Nov 24 00:27:00.979778 kernel: Demotion targets for Node 0: null Nov 24 00:27:00.979786 kernel: Key type .fscrypt registered Nov 24 00:27:00.979793 kernel: Key type fscrypt-provisioning registered Nov 24 00:27:00.979801 kernel: ima: No TPM chip found, activating TPM-bypass! Nov 24 00:27:00.979812 kernel: ima: Allocated hash algorithm: sha1 Nov 24 00:27:00.979819 kernel: ima: No architecture policies found Nov 24 00:27:00.979827 kernel: clk: Disabling unused clocks Nov 24 00:27:00.979835 kernel: Warning: unable to open an initial console. Nov 24 00:27:00.979843 kernel: Freeing unused kernel image (initmem) memory: 46188K Nov 24 00:27:00.979851 kernel: Write protecting the kernel read-only data: 40960k Nov 24 00:27:00.979860 kernel: Freeing unused kernel image (rodata/data gap) memory: 560K Nov 24 00:27:00.979867 kernel: Run /init as init process Nov 24 00:27:00.979875 kernel: with arguments: Nov 24 00:27:00.979887 kernel: /init Nov 24 00:27:00.979895 kernel: with environment: Nov 24 00:27:00.979923 kernel: HOME=/ Nov 24 00:27:00.979933 kernel: TERM=linux Nov 24 00:27:00.979943 systemd[1]: Successfully made /usr/ read-only. Nov 24 00:27:00.979957 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Nov 24 00:27:00.979966 systemd[1]: Detected virtualization kvm. Nov 24 00:27:00.979978 systemd[1]: Detected architecture x86-64. Nov 24 00:27:00.979986 systemd[1]: Running in initrd. Nov 24 00:27:00.979993 systemd[1]: No hostname configured, using default hostname. Nov 24 00:27:00.980002 systemd[1]: Hostname set to . Nov 24 00:27:00.980010 systemd[1]: Initializing machine ID from random generator. Nov 24 00:27:00.980018 systemd[1]: Queued start job for default target initrd.target. Nov 24 00:27:00.980026 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 24 00:27:00.980035 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 24 00:27:00.980046 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Nov 24 00:27:00.980055 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 24 00:27:00.980063 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Nov 24 00:27:00.980072 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Nov 24 00:27:00.980081 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Nov 24 00:27:00.980090 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Nov 24 00:27:00.980098 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 24 00:27:00.980109 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 24 00:27:00.980117 systemd[1]: Reached target paths.target - Path Units. Nov 24 00:27:00.980125 systemd[1]: Reached target slices.target - Slice Units. Nov 24 00:27:00.980133 systemd[1]: Reached target swap.target - Swaps. Nov 24 00:27:00.980142 systemd[1]: Reached target timers.target - Timer Units. Nov 24 00:27:00.980150 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Nov 24 00:27:00.980158 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 24 00:27:00.980166 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Nov 24 00:27:00.980174 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Nov 24 00:27:00.980185 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 24 00:27:00.980197 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 24 00:27:00.980208 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 24 00:27:00.980216 systemd[1]: Reached target sockets.target - Socket Units. Nov 24 00:27:00.980224 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Nov 24 00:27:00.980236 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 24 00:27:00.980244 systemd[1]: Finished network-cleanup.service - Network Cleanup. Nov 24 00:27:00.980252 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Nov 24 00:27:00.980261 systemd[1]: Starting systemd-fsck-usr.service... Nov 24 00:27:00.980269 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 24 00:27:00.980277 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 24 00:27:00.980285 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 24 00:27:00.980293 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Nov 24 00:27:00.980305 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 24 00:27:00.980350 systemd-journald[187]: Collecting audit messages is disabled. Nov 24 00:27:00.980563 systemd[1]: Finished systemd-fsck-usr.service. Nov 24 00:27:00.980572 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 24 00:27:00.980582 systemd-journald[187]: Journal started Nov 24 00:27:00.980601 systemd-journald[187]: Runtime Journal (/run/log/journal/3882bbbbefd64b15847a8885cd9ef2fc) is 8M, max 78.2M, 70.2M free. Nov 24 00:27:00.946865 systemd-modules-load[188]: Inserted module 'overlay' Nov 24 00:27:00.990742 systemd[1]: Started systemd-journald.service - Journal Service. Nov 24 00:27:00.992751 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 24 00:27:00.994295 systemd-modules-load[188]: Inserted module 'br_netfilter' Nov 24 00:27:01.105742 kernel: Bridge firewalling registered Nov 24 00:27:01.110925 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 24 00:27:01.113958 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:27:01.119867 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 24 00:27:01.123848 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 24 00:27:01.140496 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 24 00:27:01.143335 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 24 00:27:01.146850 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 24 00:27:01.155294 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 24 00:27:01.165222 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 24 00:27:01.167894 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Nov 24 00:27:01.170421 systemd-tmpfiles[206]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Nov 24 00:27:01.170769 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 24 00:27:01.180915 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 24 00:27:01.184858 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 24 00:27:01.201739 dracut-cmdline[223]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=1969a6ee0c0ec5507eb68849c160e94c58e52d2291c767873af68a1f52b30801 Nov 24 00:27:01.234874 systemd-resolved[226]: Positive Trust Anchors: Nov 24 00:27:01.234897 systemd-resolved[226]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 24 00:27:01.234931 systemd-resolved[226]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 24 00:27:01.241994 systemd-resolved[226]: Defaulting to hostname 'linux'. Nov 24 00:27:01.243189 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 24 00:27:01.244370 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 24 00:27:01.297761 kernel: SCSI subsystem initialized Nov 24 00:27:01.307741 kernel: Loading iSCSI transport class v2.0-870. Nov 24 00:27:01.318772 kernel: iscsi: registered transport (tcp) Nov 24 00:27:01.341695 kernel: iscsi: registered transport (qla4xxx) Nov 24 00:27:01.341763 kernel: QLogic iSCSI HBA Driver Nov 24 00:27:01.362072 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 24 00:27:01.376065 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 24 00:27:01.379023 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 24 00:27:01.422853 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Nov 24 00:27:01.426172 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Nov 24 00:27:01.488768 kernel: raid6: avx2x4 gen() 30958 MB/s Nov 24 00:27:01.506750 kernel: raid6: avx2x2 gen() 24928 MB/s Nov 24 00:27:01.524827 kernel: raid6: avx2x1 gen() 17493 MB/s Nov 24 00:27:01.524862 kernel: raid6: using algorithm avx2x4 gen() 30958 MB/s Nov 24 00:27:01.545024 kernel: raid6: .... xor() 2414 MB/s, rmw enabled Nov 24 00:27:01.545054 kernel: raid6: using avx2x2 recovery algorithm Nov 24 00:27:01.568760 kernel: xor: automatically using best checksumming function avx Nov 24 00:27:01.715771 kernel: Btrfs loaded, zoned=no, fsverity=no Nov 24 00:27:01.723362 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Nov 24 00:27:01.726305 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 24 00:27:01.750204 systemd-udevd[435]: Using default interface naming scheme 'v255'. Nov 24 00:27:01.756292 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 24 00:27:01.760740 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Nov 24 00:27:01.780961 dracut-pre-trigger[442]: rd.md=0: removing MD RAID activation Nov 24 00:27:01.809208 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Nov 24 00:27:01.811578 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 24 00:27:01.888136 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 24 00:27:01.893170 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Nov 24 00:27:01.944754 kernel: virtio_scsi virtio0: 2/0/0 default/read/poll queues Nov 24 00:27:01.949750 kernel: scsi host0: Virtio SCSI HBA Nov 24 00:27:02.142825 kernel: cryptd: max_cpu_qlen set to 1000 Nov 24 00:27:02.151754 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Nov 24 00:27:02.160744 kernel: libata version 3.00 loaded. Nov 24 00:27:02.174561 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 24 00:27:02.174859 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:27:02.203318 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Nov 24 00:27:02.207080 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 24 00:27:02.211964 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Nov 24 00:27:02.224187 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Nov 24 00:27:02.226912 kernel: ahci 0000:00:1f.2: version 3.0 Nov 24 00:27:02.227130 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Nov 24 00:27:02.235747 kernel: AES CTR mode by8 optimization enabled Nov 24 00:27:02.235775 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Nov 24 00:27:02.252445 kernel: sd 0:0:0:0: Power-on or device reset occurred Nov 24 00:27:02.252638 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Nov 24 00:27:02.253068 kernel: sd 0:0:0:0: [sda] 167739392 512-byte logical blocks: (85.9 GB/80.0 GiB) Nov 24 00:27:02.253227 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Nov 24 00:27:02.253376 kernel: sd 0:0:0:0: [sda] Write Protect is off Nov 24 00:27:02.253526 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Nov 24 00:27:02.260734 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Nov 24 00:27:02.260915 kernel: scsi host1: ahci Nov 24 00:27:02.266738 kernel: scsi host2: ahci Nov 24 00:27:02.281140 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Nov 24 00:27:02.281163 kernel: GPT:9289727 != 167739391 Nov 24 00:27:02.281175 kernel: GPT:Alternate GPT header not at the end of the disk. Nov 24 00:27:02.281191 kernel: GPT:9289727 != 167739391 Nov 24 00:27:02.281201 kernel: GPT: Use GNU Parted to correct GPT errors. Nov 24 00:27:02.281210 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 24 00:27:02.289308 kernel: scsi host3: ahci Nov 24 00:27:02.289349 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Nov 24 00:27:02.292735 kernel: scsi host4: ahci Nov 24 00:27:02.297737 kernel: scsi host5: ahci Nov 24 00:27:02.297925 kernel: scsi host6: ahci Nov 24 00:27:02.299816 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3100 irq 46 lpm-pol 1 Nov 24 00:27:02.299840 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3180 irq 46 lpm-pol 1 Nov 24 00:27:02.299851 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3200 irq 46 lpm-pol 1 Nov 24 00:27:02.299862 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3280 irq 46 lpm-pol 1 Nov 24 00:27:02.299872 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3300 irq 46 lpm-pol 1 Nov 24 00:27:02.299881 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3380 irq 46 lpm-pol 1 Nov 24 00:27:02.386073 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Nov 24 00:27:02.465315 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:27:02.476026 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Nov 24 00:27:02.485551 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Nov 24 00:27:02.493377 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Nov 24 00:27:02.494176 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Nov 24 00:27:02.498345 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Nov 24 00:27:02.526973 disk-uuid[599]: Primary Header is updated. Nov 24 00:27:02.526973 disk-uuid[599]: Secondary Entries is updated. Nov 24 00:27:02.526973 disk-uuid[599]: Secondary Header is updated. Nov 24 00:27:02.539757 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 24 00:27:02.555737 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 24 00:27:02.615908 kernel: ata5: SATA link down (SStatus 0 SControl 300) Nov 24 00:27:02.615936 kernel: ata6: SATA link down (SStatus 0 SControl 300) Nov 24 00:27:02.615948 kernel: ata1: SATA link down (SStatus 0 SControl 300) Nov 24 00:27:02.617486 kernel: ata2: SATA link down (SStatus 0 SControl 300) Nov 24 00:27:02.621143 kernel: ata3: SATA link down (SStatus 0 SControl 300) Nov 24 00:27:02.621727 kernel: ata4: SATA link down (SStatus 0 SControl 300) Nov 24 00:27:02.735945 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Nov 24 00:27:02.761307 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Nov 24 00:27:02.762195 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 24 00:27:02.763944 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 24 00:27:02.766675 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Nov 24 00:27:02.791880 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Nov 24 00:27:03.558791 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 24 00:27:03.559388 disk-uuid[600]: The operation has completed successfully. Nov 24 00:27:03.614964 systemd[1]: disk-uuid.service: Deactivated successfully. Nov 24 00:27:03.615110 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Nov 24 00:27:03.640450 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Nov 24 00:27:03.652568 sh[635]: Success Nov 24 00:27:03.679655 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Nov 24 00:27:03.679689 kernel: device-mapper: uevent: version 1.0.3 Nov 24 00:27:03.683749 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Nov 24 00:27:03.694800 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Nov 24 00:27:03.738903 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Nov 24 00:27:03.743785 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Nov 24 00:27:03.760067 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Nov 24 00:27:03.772764 kernel: BTRFS: device fsid c993ebd2-0e38-4cfc-8615-2c75294bea72 devid 1 transid 36 /dev/mapper/usr (254:0) scanned by mount (647) Nov 24 00:27:03.779959 kernel: BTRFS info (device dm-0): first mount of filesystem c993ebd2-0e38-4cfc-8615-2c75294bea72 Nov 24 00:27:03.779989 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Nov 24 00:27:03.792504 kernel: BTRFS info (device dm-0): enabling ssd optimizations Nov 24 00:27:03.792530 kernel: BTRFS info (device dm-0): disabling log replay at mount time Nov 24 00:27:03.792544 kernel: BTRFS info (device dm-0): enabling free space tree Nov 24 00:27:03.797071 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Nov 24 00:27:03.799165 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Nov 24 00:27:03.801036 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Nov 24 00:27:03.802884 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Nov 24 00:27:03.805815 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Nov 24 00:27:03.837754 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (683) Nov 24 00:27:03.844571 kernel: BTRFS info (device sda6): first mount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:27:03.844603 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 24 00:27:03.854645 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 24 00:27:03.854673 kernel: BTRFS info (device sda6): turning on async discard Nov 24 00:27:03.854693 kernel: BTRFS info (device sda6): enabling free space tree Nov 24 00:27:03.863746 kernel: BTRFS info (device sda6): last unmount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:27:03.866108 systemd[1]: Finished ignition-setup.service - Ignition (setup). Nov 24 00:27:03.870606 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Nov 24 00:27:03.968990 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 24 00:27:03.974020 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 24 00:27:03.993105 ignition[755]: Ignition 2.22.0 Nov 24 00:27:03.993186 ignition[755]: Stage: fetch-offline Nov 24 00:27:03.993224 ignition[755]: no configs at "/usr/lib/ignition/base.d" Nov 24 00:27:03.993234 ignition[755]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Nov 24 00:27:03.993317 ignition[755]: parsed url from cmdline: "" Nov 24 00:27:04.000217 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Nov 24 00:27:03.993321 ignition[755]: no config URL provided Nov 24 00:27:03.993327 ignition[755]: reading system config file "/usr/lib/ignition/user.ign" Nov 24 00:27:03.993335 ignition[755]: no config at "/usr/lib/ignition/user.ign" Nov 24 00:27:03.993341 ignition[755]: failed to fetch config: resource requires networking Nov 24 00:27:03.995129 ignition[755]: Ignition finished successfully Nov 24 00:27:04.020090 systemd-networkd[820]: lo: Link UP Nov 24 00:27:04.020106 systemd-networkd[820]: lo: Gained carrier Nov 24 00:27:04.021812 systemd-networkd[820]: Enumeration completed Nov 24 00:27:04.022235 systemd-networkd[820]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 24 00:27:04.022241 systemd-networkd[820]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 24 00:27:04.023016 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 24 00:27:04.024205 systemd-networkd[820]: eth0: Link UP Nov 24 00:27:04.024389 systemd-networkd[820]: eth0: Gained carrier Nov 24 00:27:04.024399 systemd-networkd[820]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 24 00:27:04.024875 systemd[1]: Reached target network.target - Network. Nov 24 00:27:04.030975 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Nov 24 00:27:04.076634 ignition[825]: Ignition 2.22.0 Nov 24 00:27:04.076648 ignition[825]: Stage: fetch Nov 24 00:27:04.077781 ignition[825]: no configs at "/usr/lib/ignition/base.d" Nov 24 00:27:04.077793 ignition[825]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Nov 24 00:27:04.077894 ignition[825]: parsed url from cmdline: "" Nov 24 00:27:04.077899 ignition[825]: no config URL provided Nov 24 00:27:04.077905 ignition[825]: reading system config file "/usr/lib/ignition/user.ign" Nov 24 00:27:04.077914 ignition[825]: no config at "/usr/lib/ignition/user.ign" Nov 24 00:27:04.077943 ignition[825]: PUT http://169.254.169.254/v1/token: attempt #1 Nov 24 00:27:04.078454 ignition[825]: PUT error: Put "http://169.254.169.254/v1/token": dial tcp 169.254.169.254:80: connect: network is unreachable Nov 24 00:27:04.279218 ignition[825]: PUT http://169.254.169.254/v1/token: attempt #2 Nov 24 00:27:04.279471 ignition[825]: PUT error: Put "http://169.254.169.254/v1/token": dial tcp 169.254.169.254:80: connect: network is unreachable Nov 24 00:27:04.680128 ignition[825]: PUT http://169.254.169.254/v1/token: attempt #3 Nov 24 00:27:04.680384 ignition[825]: PUT error: Put "http://169.254.169.254/v1/token": dial tcp 169.254.169.254:80: connect: network is unreachable Nov 24 00:27:04.812793 systemd-networkd[820]: eth0: DHCPv4 address 172.237.154.236/24, gateway 172.237.154.1 acquired from 23.205.167.131 Nov 24 00:27:05.480590 ignition[825]: PUT http://169.254.169.254/v1/token: attempt #4 Nov 24 00:27:05.573556 ignition[825]: PUT result: OK Nov 24 00:27:05.573646 ignition[825]: GET http://169.254.169.254/v1/user-data: attempt #1 Nov 24 00:27:05.688802 ignition[825]: GET result: OK Nov 24 00:27:05.688931 ignition[825]: parsing config with SHA512: abe565fe6479a489cafc68f396b9a5d6fd316177105630ef17bded4462c574d6dec62668fcca6efb4f8d62d94b09f92e75e3b0a96fdd25c4ab46142e859ccef2 Nov 24 00:27:05.692377 unknown[825]: fetched base config from "system" Nov 24 00:27:05.692386 unknown[825]: fetched base config from "system" Nov 24 00:27:05.692611 ignition[825]: fetch: fetch complete Nov 24 00:27:05.692392 unknown[825]: fetched user config from "akamai" Nov 24 00:27:05.692617 ignition[825]: fetch: fetch passed Nov 24 00:27:05.697171 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Nov 24 00:27:05.692660 ignition[825]: Ignition finished successfully Nov 24 00:27:05.719658 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Nov 24 00:27:05.760085 ignition[833]: Ignition 2.22.0 Nov 24 00:27:05.760104 ignition[833]: Stage: kargs Nov 24 00:27:05.760252 ignition[833]: no configs at "/usr/lib/ignition/base.d" Nov 24 00:27:05.760263 ignition[833]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Nov 24 00:27:05.761018 ignition[833]: kargs: kargs passed Nov 24 00:27:05.761061 ignition[833]: Ignition finished successfully Nov 24 00:27:05.766129 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Nov 24 00:27:05.768897 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Nov 24 00:27:05.796268 ignition[839]: Ignition 2.22.0 Nov 24 00:27:05.796285 ignition[839]: Stage: disks Nov 24 00:27:05.796431 ignition[839]: no configs at "/usr/lib/ignition/base.d" Nov 24 00:27:05.796442 ignition[839]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Nov 24 00:27:05.797201 ignition[839]: disks: disks passed Nov 24 00:27:05.799082 systemd[1]: Finished ignition-disks.service - Ignition (disks). Nov 24 00:27:05.797254 ignition[839]: Ignition finished successfully Nov 24 00:27:05.801021 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Nov 24 00:27:05.802687 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Nov 24 00:27:05.804246 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 24 00:27:05.806003 systemd[1]: Reached target sysinit.target - System Initialization. Nov 24 00:27:05.807738 systemd[1]: Reached target basic.target - Basic System. Nov 24 00:27:05.808902 systemd-networkd[820]: eth0: Gained IPv6LL Nov 24 00:27:05.811840 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Nov 24 00:27:05.847659 systemd-fsck[848]: ROOT: clean, 15/553520 files, 52789/553472 blocks Nov 24 00:27:05.850915 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Nov 24 00:27:05.853801 systemd[1]: Mounting sysroot.mount - /sysroot... Nov 24 00:27:05.968740 kernel: EXT4-fs (sda9): mounted filesystem 5d9d0447-100f-4769-adb5-76fdba966eb2 r/w with ordered data mode. Quota mode: none. Nov 24 00:27:05.969653 systemd[1]: Mounted sysroot.mount - /sysroot. Nov 24 00:27:05.970932 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Nov 24 00:27:05.973282 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 24 00:27:05.976775 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Nov 24 00:27:05.980026 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Nov 24 00:27:05.980869 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Nov 24 00:27:05.980908 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Nov 24 00:27:05.986963 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Nov 24 00:27:05.990857 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Nov 24 00:27:05.998897 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (856) Nov 24 00:27:06.003286 kernel: BTRFS info (device sda6): first mount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:27:06.003318 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 24 00:27:06.012010 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 24 00:27:06.012039 kernel: BTRFS info (device sda6): turning on async discard Nov 24 00:27:06.013749 kernel: BTRFS info (device sda6): enabling free space tree Nov 24 00:27:06.019483 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 24 00:27:06.064436 initrd-setup-root[880]: cut: /sysroot/etc/passwd: No such file or directory Nov 24 00:27:06.071346 initrd-setup-root[887]: cut: /sysroot/etc/group: No such file or directory Nov 24 00:27:06.077440 initrd-setup-root[894]: cut: /sysroot/etc/shadow: No such file or directory Nov 24 00:27:06.083774 initrd-setup-root[901]: cut: /sysroot/etc/gshadow: No such file or directory Nov 24 00:27:06.200081 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Nov 24 00:27:06.204002 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Nov 24 00:27:06.206352 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Nov 24 00:27:06.224375 systemd[1]: sysroot-oem.mount: Deactivated successfully. Nov 24 00:27:06.231762 kernel: BTRFS info (device sda6): last unmount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:27:06.241237 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Nov 24 00:27:06.271353 ignition[972]: INFO : Ignition 2.22.0 Nov 24 00:27:06.273813 ignition[972]: INFO : Stage: mount Nov 24 00:27:06.273813 ignition[972]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 24 00:27:06.273813 ignition[972]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Nov 24 00:27:06.273813 ignition[972]: INFO : mount: mount passed Nov 24 00:27:06.273813 ignition[972]: INFO : Ignition finished successfully Nov 24 00:27:06.277455 systemd[1]: Finished ignition-mount.service - Ignition (mount). Nov 24 00:27:06.281954 systemd[1]: Starting ignition-files.service - Ignition (files)... Nov 24 00:27:06.971434 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 24 00:27:07.002755 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (981) Nov 24 00:27:07.002794 kernel: BTRFS info (device sda6): first mount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:27:07.005996 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 24 00:27:07.014902 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 24 00:27:07.014927 kernel: BTRFS info (device sda6): turning on async discard Nov 24 00:27:07.019158 kernel: BTRFS info (device sda6): enabling free space tree Nov 24 00:27:07.021797 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 24 00:27:07.058516 ignition[997]: INFO : Ignition 2.22.0 Nov 24 00:27:07.058516 ignition[997]: INFO : Stage: files Nov 24 00:27:07.060361 ignition[997]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 24 00:27:07.060361 ignition[997]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Nov 24 00:27:07.060361 ignition[997]: DEBUG : files: compiled without relabeling support, skipping Nov 24 00:27:07.063234 ignition[997]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Nov 24 00:27:07.063234 ignition[997]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Nov 24 00:27:07.065472 ignition[997]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Nov 24 00:27:07.065472 ignition[997]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Nov 24 00:27:07.067667 ignition[997]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Nov 24 00:27:07.066191 unknown[997]: wrote ssh authorized keys file for user: core Nov 24 00:27:07.069496 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 24 00:27:07.069496 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Nov 24 00:27:07.382761 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Nov 24 00:27:07.584617 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 24 00:27:07.584617 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Nov 24 00:27:07.587462 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Nov 24 00:27:07.587462 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Nov 24 00:27:07.587462 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Nov 24 00:27:07.587462 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 24 00:27:07.587462 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 24 00:27:07.587462 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 24 00:27:07.587462 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 24 00:27:07.587462 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Nov 24 00:27:07.596481 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Nov 24 00:27:07.596481 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 24 00:27:07.596481 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 24 00:27:07.596481 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 24 00:27:07.596481 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Nov 24 00:27:07.919422 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Nov 24 00:27:08.454445 ignition[997]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 24 00:27:08.454445 ignition[997]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Nov 24 00:27:08.457651 ignition[997]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 24 00:27:08.459355 ignition[997]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 24 00:27:08.459355 ignition[997]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Nov 24 00:27:08.459355 ignition[997]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Nov 24 00:27:08.464017 ignition[997]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Nov 24 00:27:08.464017 ignition[997]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Nov 24 00:27:08.464017 ignition[997]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Nov 24 00:27:08.464017 ignition[997]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Nov 24 00:27:08.464017 ignition[997]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Nov 24 00:27:08.464017 ignition[997]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Nov 24 00:27:08.464017 ignition[997]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Nov 24 00:27:08.464017 ignition[997]: INFO : files: files passed Nov 24 00:27:08.464017 ignition[997]: INFO : Ignition finished successfully Nov 24 00:27:08.467081 systemd[1]: Finished ignition-files.service - Ignition (files). Nov 24 00:27:08.471855 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Nov 24 00:27:08.477031 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Nov 24 00:27:08.485366 systemd[1]: ignition-quench.service: Deactivated successfully. Nov 24 00:27:08.495867 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Nov 24 00:27:08.507809 initrd-setup-root-after-ignition[1028]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 24 00:27:08.507809 initrd-setup-root-after-ignition[1028]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Nov 24 00:27:08.511208 initrd-setup-root-after-ignition[1032]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 24 00:27:08.513995 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 24 00:27:08.516463 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Nov 24 00:27:08.519353 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Nov 24 00:27:08.566007 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Nov 24 00:27:08.566143 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Nov 24 00:27:08.567896 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Nov 24 00:27:08.569255 systemd[1]: Reached target initrd.target - Initrd Default Target. Nov 24 00:27:08.571029 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Nov 24 00:27:08.572776 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Nov 24 00:27:08.614065 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 24 00:27:08.616824 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Nov 24 00:27:08.636198 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Nov 24 00:27:08.637451 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 24 00:27:08.639145 systemd[1]: Stopped target timers.target - Timer Units. Nov 24 00:27:08.640748 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Nov 24 00:27:08.640889 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 24 00:27:08.642876 systemd[1]: Stopped target initrd.target - Initrd Default Target. Nov 24 00:27:08.644011 systemd[1]: Stopped target basic.target - Basic System. Nov 24 00:27:08.645625 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Nov 24 00:27:08.647115 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Nov 24 00:27:08.648597 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Nov 24 00:27:08.650211 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Nov 24 00:27:08.651925 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Nov 24 00:27:08.653536 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Nov 24 00:27:08.655239 systemd[1]: Stopped target sysinit.target - System Initialization. Nov 24 00:27:08.656882 systemd[1]: Stopped target local-fs.target - Local File Systems. Nov 24 00:27:08.658470 systemd[1]: Stopped target swap.target - Swaps. Nov 24 00:27:08.660059 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Nov 24 00:27:08.660196 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Nov 24 00:27:08.661966 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Nov 24 00:27:08.663087 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 24 00:27:08.664513 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Nov 24 00:27:08.667017 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 24 00:27:08.667967 systemd[1]: dracut-initqueue.service: Deactivated successfully. Nov 24 00:27:08.668064 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Nov 24 00:27:08.670147 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Nov 24 00:27:08.670303 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 24 00:27:08.671310 systemd[1]: ignition-files.service: Deactivated successfully. Nov 24 00:27:08.671445 systemd[1]: Stopped ignition-files.service - Ignition (files). Nov 24 00:27:08.674798 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Nov 24 00:27:08.677885 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Nov 24 00:27:08.678575 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Nov 24 00:27:08.678686 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Nov 24 00:27:08.681154 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Nov 24 00:27:08.681293 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Nov 24 00:27:08.690949 systemd[1]: initrd-cleanup.service: Deactivated successfully. Nov 24 00:27:08.691061 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Nov 24 00:27:08.711241 ignition[1052]: INFO : Ignition 2.22.0 Nov 24 00:27:08.734465 ignition[1052]: INFO : Stage: umount Nov 24 00:27:08.734465 ignition[1052]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 24 00:27:08.734465 ignition[1052]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Nov 24 00:27:08.734465 ignition[1052]: INFO : umount: umount passed Nov 24 00:27:08.734465 ignition[1052]: INFO : Ignition finished successfully Nov 24 00:27:08.711373 systemd[1]: sysroot-boot.mount: Deactivated successfully. Nov 24 00:27:08.719237 systemd[1]: ignition-mount.service: Deactivated successfully. Nov 24 00:27:08.719347 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Nov 24 00:27:08.735498 systemd[1]: sysroot-boot.service: Deactivated successfully. Nov 24 00:27:08.735648 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Nov 24 00:27:08.737450 systemd[1]: ignition-disks.service: Deactivated successfully. Nov 24 00:27:08.737540 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Nov 24 00:27:08.739411 systemd[1]: ignition-kargs.service: Deactivated successfully. Nov 24 00:27:08.739463 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Nov 24 00:27:08.741008 systemd[1]: ignition-fetch.service: Deactivated successfully. Nov 24 00:27:08.741056 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Nov 24 00:27:08.742424 systemd[1]: Stopped target network.target - Network. Nov 24 00:27:08.743841 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Nov 24 00:27:08.743899 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Nov 24 00:27:08.745343 systemd[1]: Stopped target paths.target - Path Units. Nov 24 00:27:08.746792 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Nov 24 00:27:08.750748 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 24 00:27:08.752160 systemd[1]: Stopped target slices.target - Slice Units. Nov 24 00:27:08.753784 systemd[1]: Stopped target sockets.target - Socket Units. Nov 24 00:27:08.755245 systemd[1]: iscsid.socket: Deactivated successfully. Nov 24 00:27:08.755302 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Nov 24 00:27:08.756746 systemd[1]: iscsiuio.socket: Deactivated successfully. Nov 24 00:27:08.756792 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 24 00:27:08.758673 systemd[1]: ignition-setup.service: Deactivated successfully. Nov 24 00:27:08.758745 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Nov 24 00:27:08.760237 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Nov 24 00:27:08.760287 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Nov 24 00:27:08.761860 systemd[1]: initrd-setup-root.service: Deactivated successfully. Nov 24 00:27:08.761914 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Nov 24 00:27:08.763792 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Nov 24 00:27:08.765682 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Nov 24 00:27:08.773127 systemd[1]: systemd-resolved.service: Deactivated successfully. Nov 24 00:27:08.773260 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Nov 24 00:27:08.778005 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Nov 24 00:27:08.778261 systemd[1]: systemd-networkd.service: Deactivated successfully. Nov 24 00:27:08.778394 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Nov 24 00:27:08.780862 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Nov 24 00:27:08.781348 systemd[1]: Stopped target network-pre.target - Preparation for Network. Nov 24 00:27:08.783161 systemd[1]: systemd-networkd.socket: Deactivated successfully. Nov 24 00:27:08.783204 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Nov 24 00:27:08.785652 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Nov 24 00:27:08.787153 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Nov 24 00:27:08.787210 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 24 00:27:08.789296 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 24 00:27:08.789373 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 24 00:27:08.790915 systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 24 00:27:08.790970 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Nov 24 00:27:08.792885 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Nov 24 00:27:08.792938 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 24 00:27:08.796851 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 24 00:27:08.801437 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Nov 24 00:27:08.801506 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Nov 24 00:27:08.816200 systemd[1]: network-cleanup.service: Deactivated successfully. Nov 24 00:27:08.816347 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Nov 24 00:27:08.820229 systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 24 00:27:08.820430 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 24 00:27:08.822539 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Nov 24 00:27:08.822615 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Nov 24 00:27:08.824099 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Nov 24 00:27:08.824140 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Nov 24 00:27:08.825908 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Nov 24 00:27:08.825960 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Nov 24 00:27:08.828278 systemd[1]: dracut-cmdline.service: Deactivated successfully. Nov 24 00:27:08.828332 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Nov 24 00:27:08.829902 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 24 00:27:08.829956 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 24 00:27:08.832813 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Nov 24 00:27:08.835170 systemd[1]: systemd-network-generator.service: Deactivated successfully. Nov 24 00:27:08.835225 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Nov 24 00:27:08.837160 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Nov 24 00:27:08.837212 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 24 00:27:08.839165 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Nov 24 00:27:08.839217 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 24 00:27:08.840936 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Nov 24 00:27:08.840983 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Nov 24 00:27:08.842034 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 24 00:27:08.842085 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:27:08.847018 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Nov 24 00:27:08.847082 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Nov 24 00:27:08.847129 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Nov 24 00:27:08.847178 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Nov 24 00:27:08.849035 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Nov 24 00:27:08.849136 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Nov 24 00:27:08.850742 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Nov 24 00:27:08.852791 systemd[1]: Starting initrd-switch-root.service - Switch Root... Nov 24 00:27:08.878918 systemd[1]: Switching root. Nov 24 00:27:08.916010 systemd-journald[187]: Journal stopped Nov 24 00:27:10.148896 systemd-journald[187]: Received SIGTERM from PID 1 (systemd). Nov 24 00:27:10.148931 kernel: SELinux: policy capability network_peer_controls=1 Nov 24 00:27:10.148944 kernel: SELinux: policy capability open_perms=1 Nov 24 00:27:10.148954 kernel: SELinux: policy capability extended_socket_class=1 Nov 24 00:27:10.148963 kernel: SELinux: policy capability always_check_network=0 Nov 24 00:27:10.148975 kernel: SELinux: policy capability cgroup_seclabel=1 Nov 24 00:27:10.148986 kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 24 00:27:10.148996 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Nov 24 00:27:10.149006 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Nov 24 00:27:10.149015 kernel: SELinux: policy capability userspace_initial_context=0 Nov 24 00:27:10.149025 kernel: audit: type=1403 audit(1763944029.102:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Nov 24 00:27:10.149036 systemd[1]: Successfully loaded SELinux policy in 79.340ms. Nov 24 00:27:10.149050 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 8ms. Nov 24 00:27:10.149062 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Nov 24 00:27:10.149079 systemd[1]: Detected virtualization kvm. Nov 24 00:27:10.149090 systemd[1]: Detected architecture x86-64. Nov 24 00:27:10.149108 systemd[1]: Detected first boot. Nov 24 00:27:10.149119 systemd[1]: Initializing machine ID from random generator. Nov 24 00:27:10.149130 zram_generator::config[1095]: No configuration found. Nov 24 00:27:10.149141 kernel: Guest personality initialized and is inactive Nov 24 00:27:10.149151 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Nov 24 00:27:10.149161 kernel: Initialized host personality Nov 24 00:27:10.149171 kernel: NET: Registered PF_VSOCK protocol family Nov 24 00:27:10.149182 systemd[1]: Populated /etc with preset unit settings. Nov 24 00:27:10.149197 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Nov 24 00:27:10.149208 systemd[1]: initrd-switch-root.service: Deactivated successfully. Nov 24 00:27:10.149219 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Nov 24 00:27:10.149229 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Nov 24 00:27:10.149240 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Nov 24 00:27:10.149250 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Nov 24 00:27:10.149261 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Nov 24 00:27:10.149275 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Nov 24 00:27:10.149286 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Nov 24 00:27:10.149297 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Nov 24 00:27:10.149307 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Nov 24 00:27:10.149318 systemd[1]: Created slice user.slice - User and Session Slice. Nov 24 00:27:10.149328 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 24 00:27:10.149339 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 24 00:27:10.149350 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Nov 24 00:27:10.149363 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Nov 24 00:27:10.149377 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Nov 24 00:27:10.149388 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 24 00:27:10.149399 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Nov 24 00:27:10.149410 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 24 00:27:10.149421 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 24 00:27:10.149433 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Nov 24 00:27:10.149447 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Nov 24 00:27:10.149458 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Nov 24 00:27:10.149469 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Nov 24 00:27:10.149480 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 24 00:27:10.149491 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 24 00:27:10.149502 systemd[1]: Reached target slices.target - Slice Units. Nov 24 00:27:10.149513 systemd[1]: Reached target swap.target - Swaps. Nov 24 00:27:10.151770 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Nov 24 00:27:10.151787 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Nov 24 00:27:10.151803 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Nov 24 00:27:10.151815 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 24 00:27:10.151826 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 24 00:27:10.151837 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 24 00:27:10.151851 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Nov 24 00:27:10.151862 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Nov 24 00:27:10.151873 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Nov 24 00:27:10.151884 systemd[1]: Mounting media.mount - External Media Directory... Nov 24 00:27:10.151895 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:27:10.151906 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Nov 24 00:27:10.151917 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Nov 24 00:27:10.151929 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Nov 24 00:27:10.151943 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Nov 24 00:27:10.151954 systemd[1]: Reached target machines.target - Containers. Nov 24 00:27:10.151966 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Nov 24 00:27:10.151977 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 24 00:27:10.151988 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 24 00:27:10.151999 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Nov 24 00:27:10.152010 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 24 00:27:10.152021 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 24 00:27:10.152032 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 24 00:27:10.152046 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Nov 24 00:27:10.152057 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 24 00:27:10.152068 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Nov 24 00:27:10.152079 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Nov 24 00:27:10.152090 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Nov 24 00:27:10.152101 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Nov 24 00:27:10.152112 systemd[1]: Stopped systemd-fsck-usr.service. Nov 24 00:27:10.152123 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 24 00:27:10.152137 kernel: fuse: init (API version 7.41) Nov 24 00:27:10.152148 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 24 00:27:10.152158 kernel: loop: module loaded Nov 24 00:27:10.152170 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 24 00:27:10.152181 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 24 00:27:10.152193 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Nov 24 00:27:10.152204 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Nov 24 00:27:10.152214 kernel: ACPI: bus type drm_connector registered Nov 24 00:27:10.152253 systemd-journald[1186]: Collecting audit messages is disabled. Nov 24 00:27:10.152275 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 24 00:27:10.152287 systemd[1]: verity-setup.service: Deactivated successfully. Nov 24 00:27:10.152299 systemd[1]: Stopped verity-setup.service. Nov 24 00:27:10.152313 systemd-journald[1186]: Journal started Nov 24 00:27:10.152334 systemd-journald[1186]: Runtime Journal (/run/log/journal/278a2a0f57fb42b39e914831e5d8311a) is 8M, max 78.2M, 70.2M free. Nov 24 00:27:09.757808 systemd[1]: Queued start job for default target multi-user.target. Nov 24 00:27:09.778537 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Nov 24 00:27:09.779137 systemd[1]: systemd-journald.service: Deactivated successfully. Nov 24 00:27:10.154843 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:27:10.172730 systemd[1]: Started systemd-journald.service - Journal Service. Nov 24 00:27:10.173468 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Nov 24 00:27:10.174455 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Nov 24 00:27:10.175382 systemd[1]: Mounted media.mount - External Media Directory. Nov 24 00:27:10.176301 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Nov 24 00:27:10.177320 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Nov 24 00:27:10.178263 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Nov 24 00:27:10.179386 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Nov 24 00:27:10.180596 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 24 00:27:10.181835 systemd[1]: modprobe@configfs.service: Deactivated successfully. Nov 24 00:27:10.182131 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Nov 24 00:27:10.183276 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 24 00:27:10.183565 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 24 00:27:10.184989 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 24 00:27:10.185277 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 24 00:27:10.186502 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 24 00:27:10.186793 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 24 00:27:10.190640 systemd[1]: modprobe@fuse.service: Deactivated successfully. Nov 24 00:27:10.191298 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Nov 24 00:27:10.193130 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 24 00:27:10.193444 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 24 00:27:10.194912 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 24 00:27:10.197015 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 24 00:27:10.198439 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Nov 24 00:27:10.200012 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Nov 24 00:27:10.218577 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 24 00:27:10.222856 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Nov 24 00:27:10.230786 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Nov 24 00:27:10.231542 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Nov 24 00:27:10.231575 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 24 00:27:10.233444 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Nov 24 00:27:10.240282 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Nov 24 00:27:10.243446 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 24 00:27:10.245827 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Nov 24 00:27:10.253877 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Nov 24 00:27:10.255004 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 24 00:27:10.257925 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Nov 24 00:27:10.259785 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 24 00:27:10.262485 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 24 00:27:10.264978 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Nov 24 00:27:10.270766 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 24 00:27:10.273742 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Nov 24 00:27:10.277643 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Nov 24 00:27:10.296062 systemd-journald[1186]: Time spent on flushing to /var/log/journal/278a2a0f57fb42b39e914831e5d8311a is 36.834ms for 1010 entries. Nov 24 00:27:10.296062 systemd-journald[1186]: System Journal (/var/log/journal/278a2a0f57fb42b39e914831e5d8311a) is 8M, max 195.6M, 187.6M free. Nov 24 00:27:10.365316 systemd-journald[1186]: Received client request to flush runtime journal. Nov 24 00:27:10.365370 kernel: loop0: detected capacity change from 0 to 224512 Nov 24 00:27:10.313093 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Nov 24 00:27:10.318497 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Nov 24 00:27:10.325246 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Nov 24 00:27:10.326410 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 24 00:27:10.361218 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 24 00:27:10.370126 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Nov 24 00:27:10.378658 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Nov 24 00:27:10.379624 systemd-tmpfiles[1221]: ACLs are not supported, ignoring. Nov 24 00:27:10.379636 systemd-tmpfiles[1221]: ACLs are not supported, ignoring. Nov 24 00:27:10.387696 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 24 00:27:10.390815 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Nov 24 00:27:10.394883 systemd[1]: Starting systemd-sysusers.service - Create System Users... Nov 24 00:27:10.414763 kernel: loop1: detected capacity change from 0 to 110984 Nov 24 00:27:10.452008 systemd[1]: Finished systemd-sysusers.service - Create System Users. Nov 24 00:27:10.460226 kernel: loop2: detected capacity change from 0 to 8 Nov 24 00:27:10.455257 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 24 00:27:10.483443 kernel: loop3: detected capacity change from 0 to 128560 Nov 24 00:27:10.499640 systemd-tmpfiles[1242]: ACLs are not supported, ignoring. Nov 24 00:27:10.499984 systemd-tmpfiles[1242]: ACLs are not supported, ignoring. Nov 24 00:27:10.508760 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 24 00:27:10.516743 kernel: loop4: detected capacity change from 0 to 224512 Nov 24 00:27:10.545905 kernel: loop5: detected capacity change from 0 to 110984 Nov 24 00:27:10.565824 kernel: loop6: detected capacity change from 0 to 8 Nov 24 00:27:10.576019 kernel: loop7: detected capacity change from 0 to 128560 Nov 24 00:27:10.589476 (sd-merge)[1247]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-akamai'. Nov 24 00:27:10.590158 (sd-merge)[1247]: Merged extensions into '/usr'. Nov 24 00:27:10.597637 systemd[1]: Reload requested from client PID 1220 ('systemd-sysext') (unit systemd-sysext.service)... Nov 24 00:27:10.597859 systemd[1]: Reloading... Nov 24 00:27:10.753775 zram_generator::config[1273]: No configuration found. Nov 24 00:27:10.826886 ldconfig[1215]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Nov 24 00:27:10.959106 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Nov 24 00:27:10.959655 systemd[1]: Reloading finished in 360 ms. Nov 24 00:27:10.992325 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Nov 24 00:27:10.993840 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Nov 24 00:27:10.995086 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Nov 24 00:27:11.011319 systemd[1]: Starting ensure-sysext.service... Nov 24 00:27:11.014829 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 24 00:27:11.017290 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 24 00:27:11.031118 systemd[1]: Reload requested from client PID 1317 ('systemctl') (unit ensure-sysext.service)... Nov 24 00:27:11.031139 systemd[1]: Reloading... Nov 24 00:27:11.055814 systemd-tmpfiles[1318]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Nov 24 00:27:11.056129 systemd-tmpfiles[1318]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Nov 24 00:27:11.056507 systemd-tmpfiles[1318]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Nov 24 00:27:11.056883 systemd-tmpfiles[1318]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Nov 24 00:27:11.058036 systemd-tmpfiles[1318]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Nov 24 00:27:11.058355 systemd-tmpfiles[1318]: ACLs are not supported, ignoring. Nov 24 00:27:11.058629 systemd-tmpfiles[1318]: ACLs are not supported, ignoring. Nov 24 00:27:11.067552 systemd-tmpfiles[1318]: Detected autofs mount point /boot during canonicalization of boot. Nov 24 00:27:11.067565 systemd-tmpfiles[1318]: Skipping /boot Nov 24 00:27:11.079225 systemd-udevd[1319]: Using default interface naming scheme 'v255'. Nov 24 00:27:11.088539 systemd-tmpfiles[1318]: Detected autofs mount point /boot during canonicalization of boot. Nov 24 00:27:11.088756 systemd-tmpfiles[1318]: Skipping /boot Nov 24 00:27:11.128787 zram_generator::config[1344]: No configuration found. Nov 24 00:27:11.396749 kernel: mousedev: PS/2 mouse device common for all mice Nov 24 00:27:11.426867 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Nov 24 00:27:11.430851 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Nov 24 00:27:11.443738 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Nov 24 00:27:11.443814 systemd[1]: Reloading finished in 412 ms. Nov 24 00:27:11.452727 kernel: ACPI: button: Power Button [PWRF] Nov 24 00:27:11.455289 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 24 00:27:11.457916 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 24 00:27:11.484686 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Nov 24 00:27:11.489043 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 24 00:27:11.495134 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Nov 24 00:27:11.500897 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Nov 24 00:27:11.506980 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 24 00:27:11.512878 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 24 00:27:11.516978 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Nov 24 00:27:11.525270 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:27:11.525590 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 24 00:27:11.528771 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 24 00:27:11.536215 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 24 00:27:11.544819 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 24 00:27:11.546885 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 24 00:27:11.546977 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 24 00:27:11.547057 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:27:11.550605 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:27:11.551030 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 24 00:27:11.551174 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 24 00:27:11.551245 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 24 00:27:11.551316 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:27:11.560985 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Nov 24 00:27:11.566542 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:27:11.568790 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 24 00:27:11.580905 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 24 00:27:11.582936 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 24 00:27:11.583041 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 24 00:27:11.583160 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:27:11.586882 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Nov 24 00:27:11.594172 systemd[1]: Starting systemd-update-done.service - Update is Completed... Nov 24 00:27:11.599237 systemd[1]: Finished ensure-sysext.service. Nov 24 00:27:11.606752 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Nov 24 00:27:11.617980 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Nov 24 00:27:11.626503 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Nov 24 00:27:11.627985 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Nov 24 00:27:11.639494 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 24 00:27:11.641043 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 24 00:27:11.643113 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 24 00:27:11.647059 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 24 00:27:11.650410 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 24 00:27:11.652576 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 24 00:27:11.652972 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 24 00:27:11.654367 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 24 00:27:11.655017 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 24 00:27:11.656604 systemd[1]: Finished systemd-update-done.service - Update is Completed. Nov 24 00:27:11.661627 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 24 00:27:11.673334 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 24 00:27:11.695991 augenrules[1482]: No rules Nov 24 00:27:11.709077 systemd[1]: audit-rules.service: Deactivated successfully. Nov 24 00:27:11.710050 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 24 00:27:11.741738 kernel: EDAC MC: Ver: 3.0.0 Nov 24 00:27:11.813923 systemd[1]: Started systemd-userdbd.service - User Database Manager. Nov 24 00:27:11.902148 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:27:11.911253 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Nov 24 00:27:11.916911 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Nov 24 00:27:11.944809 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Nov 24 00:27:12.013556 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Nov 24 00:27:12.014419 systemd[1]: Reached target time-set.target - System Time Set. Nov 24 00:27:12.023698 systemd-networkd[1439]: lo: Link UP Nov 24 00:27:12.023778 systemd-networkd[1439]: lo: Gained carrier Nov 24 00:27:12.025086 systemd-resolved[1440]: Positive Trust Anchors: Nov 24 00:27:12.025309 systemd-resolved[1440]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 24 00:27:12.025375 systemd-resolved[1440]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 24 00:27:12.025626 systemd-networkd[1439]: Enumeration completed Nov 24 00:27:12.025725 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 24 00:27:12.028316 systemd-networkd[1439]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 24 00:27:12.028329 systemd-networkd[1439]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 24 00:27:12.029271 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Nov 24 00:27:12.030009 systemd-resolved[1440]: Defaulting to hostname 'linux'. Nov 24 00:27:12.032821 systemd-networkd[1439]: eth0: Link UP Nov 24 00:27:12.033020 systemd-networkd[1439]: eth0: Gained carrier Nov 24 00:27:12.033044 systemd-networkd[1439]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 24 00:27:12.033250 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Nov 24 00:27:12.034855 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 24 00:27:12.036824 systemd[1]: Reached target network.target - Network. Nov 24 00:27:12.037502 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 24 00:27:12.038468 systemd[1]: Reached target sysinit.target - System Initialization. Nov 24 00:27:12.039742 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Nov 24 00:27:12.040590 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Nov 24 00:27:12.041383 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Nov 24 00:27:12.042759 systemd[1]: Started logrotate.timer - Daily rotation of log files. Nov 24 00:27:12.043671 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Nov 24 00:27:12.044445 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Nov 24 00:27:12.045220 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Nov 24 00:27:12.045259 systemd[1]: Reached target paths.target - Path Units. Nov 24 00:27:12.045965 systemd[1]: Reached target timers.target - Timer Units. Nov 24 00:27:12.047898 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Nov 24 00:27:12.050285 systemd[1]: Starting docker.socket - Docker Socket for the API... Nov 24 00:27:12.053364 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Nov 24 00:27:12.054267 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Nov 24 00:27:12.055053 systemd[1]: Reached target ssh-access.target - SSH Access Available. Nov 24 00:27:12.058382 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Nov 24 00:27:12.059433 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Nov 24 00:27:12.060894 systemd[1]: Listening on docker.socket - Docker Socket for the API. Nov 24 00:27:12.062467 systemd[1]: Reached target sockets.target - Socket Units. Nov 24 00:27:12.063177 systemd[1]: Reached target basic.target - Basic System. Nov 24 00:27:12.063923 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Nov 24 00:27:12.063955 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Nov 24 00:27:12.065529 systemd[1]: Starting containerd.service - containerd container runtime... Nov 24 00:27:12.074315 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Nov 24 00:27:12.106596 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Nov 24 00:27:12.112863 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Nov 24 00:27:12.117785 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Nov 24 00:27:12.119942 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Nov 24 00:27:12.121826 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Nov 24 00:27:12.126896 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Nov 24 00:27:12.132065 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Nov 24 00:27:12.136960 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Nov 24 00:27:12.141609 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Nov 24 00:27:12.145954 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Nov 24 00:27:12.155052 systemd[1]: Starting systemd-logind.service - User Login Management... Nov 24 00:27:12.156670 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Nov 24 00:27:12.159172 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Nov 24 00:27:12.161303 systemd[1]: Starting update-engine.service - Update Engine... Nov 24 00:27:12.167640 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Nov 24 00:27:12.172778 jq[1516]: false Nov 24 00:27:12.173898 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Nov 24 00:27:12.176742 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Refreshing passwd entry cache Nov 24 00:27:12.175792 oslogin_cache_refresh[1518]: Refreshing passwd entry cache Nov 24 00:27:12.182736 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Failure getting users, quitting Nov 24 00:27:12.182736 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Nov 24 00:27:12.182736 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Refreshing group entry cache Nov 24 00:27:12.182736 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Failure getting groups, quitting Nov 24 00:27:12.182736 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Nov 24 00:27:12.181427 oslogin_cache_refresh[1518]: Failure getting users, quitting Nov 24 00:27:12.181441 oslogin_cache_refresh[1518]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Nov 24 00:27:12.181479 oslogin_cache_refresh[1518]: Refreshing group entry cache Nov 24 00:27:12.181972 oslogin_cache_refresh[1518]: Failure getting groups, quitting Nov 24 00:27:12.181982 oslogin_cache_refresh[1518]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Nov 24 00:27:12.183951 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Nov 24 00:27:12.186081 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Nov 24 00:27:12.186345 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Nov 24 00:27:12.186903 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Nov 24 00:27:12.187138 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Nov 24 00:27:12.189265 systemd[1]: motdgen.service: Deactivated successfully. Nov 24 00:27:12.189451 extend-filesystems[1517]: Found /dev/sda6 Nov 24 00:27:12.189513 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Nov 24 00:27:12.197215 extend-filesystems[1517]: Found /dev/sda9 Nov 24 00:27:12.202869 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Nov 24 00:27:12.204129 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Nov 24 00:27:12.207483 extend-filesystems[1517]: Checking size of /dev/sda9 Nov 24 00:27:12.220753 jq[1532]: true Nov 24 00:27:12.226982 update_engine[1530]: I20251124 00:27:12.226910 1530 main.cc:92] Flatcar Update Engine starting Nov 24 00:27:12.243065 (ntainerd)[1559]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Nov 24 00:27:12.254350 extend-filesystems[1517]: Resized partition /dev/sda9 Nov 24 00:27:12.255566 jq[1555]: true Nov 24 00:27:12.256736 tar[1539]: linux-amd64/LICENSE Nov 24 00:27:12.258568 tar[1539]: linux-amd64/helm Nov 24 00:27:12.261370 extend-filesystems[1565]: resize2fs 1.47.3 (8-Jul-2025) Nov 24 00:27:12.270274 coreos-metadata[1513]: Nov 24 00:27:12.270 INFO Putting http://169.254.169.254/v1/token: Attempt #1 Nov 24 00:27:12.279668 kernel: EXT4-fs (sda9): resizing filesystem from 553472 to 20360187 blocks Nov 24 00:27:12.305891 dbus-daemon[1514]: [system] SELinux support is enabled Nov 24 00:27:12.306347 systemd[1]: Started dbus.service - D-Bus System Message Bus. Nov 24 00:27:12.311022 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Nov 24 00:27:12.311059 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Nov 24 00:27:12.312791 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Nov 24 00:27:12.312815 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Nov 24 00:27:12.321021 systemd[1]: Started update-engine.service - Update Engine. Nov 24 00:27:12.321453 update_engine[1530]: I20251124 00:27:12.321326 1530 update_check_scheduler.cc:74] Next update check in 7m3s Nov 24 00:27:12.338084 systemd-logind[1525]: Watching system buttons on /dev/input/event2 (Power Button) Nov 24 00:27:12.338119 systemd-logind[1525]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Nov 24 00:27:12.338488 systemd-logind[1525]: New seat seat0. Nov 24 00:27:12.343599 systemd[1]: Started locksmithd.service - Cluster reboot manager. Nov 24 00:27:12.344550 systemd[1]: Started systemd-logind.service - User Login Management. Nov 24 00:27:12.464009 bash[1584]: Updated "/home/core/.ssh/authorized_keys" Nov 24 00:27:12.463152 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Nov 24 00:27:12.482696 systemd[1]: Starting sshkeys.service... Nov 24 00:27:12.544075 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Nov 24 00:27:12.546774 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Nov 24 00:27:12.627237 containerd[1559]: time="2025-11-24T00:27:12Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Nov 24 00:27:12.629413 containerd[1559]: time="2025-11-24T00:27:12.629366851Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Nov 24 00:27:12.644320 locksmithd[1572]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Nov 24 00:27:12.657721 kernel: EXT4-fs (sda9): resized filesystem to 20360187 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.654980093Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="6.62µs" Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655004763Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655019793Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655166074Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655180324Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655200194Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655258384Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655268764Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655480304Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655492444Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655502364Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Nov 24 00:27:12.669763 containerd[1559]: time="2025-11-24T00:27:12.655509604Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Nov 24 00:27:12.669974 coreos-metadata[1592]: Nov 24 00:27:12.664 INFO Putting http://169.254.169.254/v1/token: Attempt #1 Nov 24 00:27:12.670191 containerd[1559]: time="2025-11-24T00:27:12.655619944Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Nov 24 00:27:12.670191 containerd[1559]: time="2025-11-24T00:27:12.655869304Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Nov 24 00:27:12.670191 containerd[1559]: time="2025-11-24T00:27:12.655897164Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Nov 24 00:27:12.670191 containerd[1559]: time="2025-11-24T00:27:12.655906154Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Nov 24 00:27:12.670191 containerd[1559]: time="2025-11-24T00:27:12.655928064Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Nov 24 00:27:12.670191 containerd[1559]: time="2025-11-24T00:27:12.656088494Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Nov 24 00:27:12.670377 containerd[1559]: time="2025-11-24T00:27:12.670344951Z" level=info msg="metadata content store policy set" policy=shared Nov 24 00:27:12.672065 extend-filesystems[1565]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Nov 24 00:27:12.672065 extend-filesystems[1565]: old_desc_blocks = 1, new_desc_blocks = 10 Nov 24 00:27:12.672065 extend-filesystems[1565]: The filesystem on /dev/sda9 is now 20360187 (4k) blocks long. Nov 24 00:27:12.685231 extend-filesystems[1517]: Resized filesystem in /dev/sda9 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675064273Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675120634Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675139864Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675191594Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675203734Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675213004Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675228884Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675239314Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675266554Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675275814Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675283984Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.675294524Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.676890014Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Nov 24 00:27:12.687354 containerd[1559]: time="2025-11-24T00:27:12.676912224Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Nov 24 00:27:12.673211 systemd[1]: extend-filesystems.service: Deactivated successfully. Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.676926914Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.676937984Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.676948254Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.676963774Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.676973704Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.676982704Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.676991764Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.677000584Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.677009374Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.677044994Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.677056424Z" level=info msg="Start snapshots syncer" Nov 24 00:27:12.688776 containerd[1559]: time="2025-11-24T00:27:12.679120386Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Nov 24 00:27:12.673477 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Nov 24 00:27:12.689033 containerd[1559]: time="2025-11-24T00:27:12.680950906Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Nov 24 00:27:12.689033 containerd[1559]: time="2025-11-24T00:27:12.681154767Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.681868937Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682410117Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682452057Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682462097Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682472167Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682484057Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682511767Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682523777Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682549317Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682558577Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682568127Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682625467Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682895757Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Nov 24 00:27:12.689135 containerd[1559]: time="2025-11-24T00:27:12.682914307Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.682930287Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.682942197Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.682952497Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.682971407Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.682989797Z" level=info msg="runtime interface created" Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.682995577Z" level=info msg="created NRI interface" Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.683005937Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.683015717Z" level=info msg="Connect containerd service" Nov 24 00:27:12.690982 containerd[1559]: time="2025-11-24T00:27:12.683036567Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Nov 24 00:27:12.692574 containerd[1559]: time="2025-11-24T00:27:12.692357252Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 24 00:27:12.786983 systemd-networkd[1439]: eth0: DHCPv4 address 172.237.154.236/24, gateway 172.237.154.1 acquired from 23.205.167.131 Nov 24 00:27:12.788415 dbus-daemon[1514]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.1' (uid=244 pid=1439 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Nov 24 00:27:12.792785 systemd-timesyncd[1458]: Network configuration changed, trying to establish connection. Nov 24 00:27:12.796020 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Nov 24 00:27:12.830345 sshd_keygen[1562]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Nov 24 00:27:12.852035 containerd[1559]: time="2025-11-24T00:27:12.852004022Z" level=info msg="Start subscribing containerd event" Nov 24 00:27:12.852698 containerd[1559]: time="2025-11-24T00:27:12.852667112Z" level=info msg="Start recovering state" Nov 24 00:27:12.852903 containerd[1559]: time="2025-11-24T00:27:12.852888752Z" level=info msg="Start event monitor" Nov 24 00:27:12.853558 containerd[1559]: time="2025-11-24T00:27:12.853198362Z" level=info msg="Start cni network conf syncer for default" Nov 24 00:27:12.853558 containerd[1559]: time="2025-11-24T00:27:12.853216322Z" level=info msg="Start streaming server" Nov 24 00:27:12.853558 containerd[1559]: time="2025-11-24T00:27:12.853224943Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Nov 24 00:27:12.853558 containerd[1559]: time="2025-11-24T00:27:12.853231763Z" level=info msg="runtime interface starting up..." Nov 24 00:27:12.853558 containerd[1559]: time="2025-11-24T00:27:12.853237953Z" level=info msg="starting plugins..." Nov 24 00:27:12.853558 containerd[1559]: time="2025-11-24T00:27:12.853253203Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Nov 24 00:27:12.855666 containerd[1559]: time="2025-11-24T00:27:12.855467814Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Nov 24 00:27:12.855666 containerd[1559]: time="2025-11-24T00:27:12.855634164Z" level=info msg=serving... address=/run/containerd/containerd.sock Nov 24 00:27:12.859449 systemd[1]: Started containerd.service - containerd container runtime. Nov 24 00:27:12.860800 containerd[1559]: time="2025-11-24T00:27:12.859473776Z" level=info msg="containerd successfully booted in 0.233877s" Nov 24 00:27:12.902676 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Nov 24 00:27:12.902951 dbus-daemon[1514]: [system] Successfully activated service 'org.freedesktop.hostname1' Nov 24 00:27:12.903365 dbus-daemon[1514]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.8' (uid=0 pid=1612 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Nov 24 00:27:12.904436 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Nov 24 00:27:12.914317 systemd[1]: Starting issuegen.service - Generate /run/issue... Nov 24 00:27:12.919954 systemd[1]: Starting polkit.service - Authorization Manager... Nov 24 00:27:12.923590 tar[1539]: linux-amd64/README.md Nov 24 00:27:12.938079 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Nov 24 00:27:12.943945 systemd[1]: issuegen.service: Deactivated successfully. Nov 24 00:27:12.944673 systemd[1]: Finished issuegen.service - Generate /run/issue. Nov 24 00:27:12.948993 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Nov 24 00:27:12.965221 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Nov 24 00:27:12.968401 systemd[1]: Started getty@tty1.service - Getty on tty1. Nov 24 00:27:12.972030 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Nov 24 00:27:12.973986 systemd[1]: Reached target getty.target - Login Prompts. Nov 24 00:27:13.006308 polkitd[1626]: Started polkitd version 126 Nov 24 00:27:13.010791 polkitd[1626]: Loading rules from directory /etc/polkit-1/rules.d Nov 24 00:27:13.011040 polkitd[1626]: Loading rules from directory /run/polkit-1/rules.d Nov 24 00:27:13.011087 polkitd[1626]: Error opening rules directory: Error opening directory “/run/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Nov 24 00:27:13.011295 polkitd[1626]: Loading rules from directory /usr/local/share/polkit-1/rules.d Nov 24 00:27:13.011322 polkitd[1626]: Error opening rules directory: Error opening directory “/usr/local/share/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Nov 24 00:27:13.011356 polkitd[1626]: Loading rules from directory /usr/share/polkit-1/rules.d Nov 24 00:27:13.012063 polkitd[1626]: Finished loading, compiling and executing 2 rules Nov 24 00:27:13.012277 systemd[1]: Started polkit.service - Authorization Manager. Nov 24 00:27:13.012538 dbus-daemon[1514]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Nov 24 00:27:13.013055 polkitd[1626]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Nov 24 00:27:13.022189 systemd-hostnamed[1612]: Hostname set to <172-237-154-236> (transient) Nov 24 00:27:13.022624 systemd-resolved[1440]: System hostname changed to '172-237-154-236'. Nov 24 00:27:13.266863 coreos-metadata[1513]: Nov 24 00:27:13.266 INFO Putting http://169.254.169.254/v1/token: Attempt #2 Nov 24 00:27:13.358748 coreos-metadata[1513]: Nov 24 00:27:13.358 INFO Fetching http://169.254.169.254/v1/instance: Attempt #1 Nov 24 00:27:13.425152 systemd-networkd[1439]: eth0: Gained IPv6LL Nov 24 00:27:13.426039 systemd-timesyncd[1458]: Network configuration changed, trying to establish connection. Nov 24 00:27:13.428088 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Nov 24 00:27:13.429471 systemd[1]: Reached target network-online.target - Network is Online. Nov 24 00:27:13.432451 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:27:13.435924 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Nov 24 00:27:13.457860 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Nov 24 00:27:13.538619 coreos-metadata[1513]: Nov 24 00:27:13.538 INFO Fetch successful Nov 24 00:27:13.538961 coreos-metadata[1513]: Nov 24 00:27:13.538 INFO Fetching http://169.254.169.254/v1/network: Attempt #1 Nov 24 00:27:13.675777 coreos-metadata[1592]: Nov 24 00:27:13.675 INFO Putting http://169.254.169.254/v1/token: Attempt #2 Nov 24 00:27:13.768565 coreos-metadata[1592]: Nov 24 00:27:13.768 INFO Fetching http://169.254.169.254/v1/ssh-keys: Attempt #1 Nov 24 00:27:13.796420 coreos-metadata[1513]: Nov 24 00:27:13.796 INFO Fetch successful Nov 24 00:27:13.903556 coreos-metadata[1592]: Nov 24 00:27:13.903 INFO Fetch successful Nov 24 00:27:13.908090 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Nov 24 00:27:13.916110 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Nov 24 00:27:13.927298 update-ssh-keys[1679]: Updated "/home/core/.ssh/authorized_keys" Nov 24 00:27:13.928802 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Nov 24 00:27:13.933123 systemd[1]: Finished sshkeys.service. Nov 24 00:27:14.007637 systemd-timesyncd[1458]: Network configuration changed, trying to establish connection. Nov 24 00:27:14.363570 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:27:14.365216 systemd[1]: Reached target multi-user.target - Multi-User System. Nov 24 00:27:14.366965 systemd[1]: Startup finished in 3.005s (kernel) + 8.416s (initrd) + 5.341s (userspace) = 16.764s. Nov 24 00:27:14.431289 (kubelet)[1689]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 24 00:27:14.996511 kubelet[1689]: E1124 00:27:14.996420 1689 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 24 00:27:15.000933 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 24 00:27:15.001206 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 24 00:27:15.002136 systemd[1]: kubelet.service: Consumed 904ms CPU time, 265.1M memory peak. Nov 24 00:27:15.377815 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Nov 24 00:27:15.379171 systemd[1]: Started sshd@0-172.237.154.236:22-139.178.68.195:33516.service - OpenSSH per-connection server daemon (139.178.68.195:33516). Nov 24 00:27:15.748650 sshd[1702]: Accepted publickey for core from 139.178.68.195 port 33516 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:27:15.750440 sshd-session[1702]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:27:15.757249 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Nov 24 00:27:15.759004 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Nov 24 00:27:15.767043 systemd-logind[1525]: New session 1 of user core. Nov 24 00:27:15.778486 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Nov 24 00:27:15.782040 systemd[1]: Starting user@500.service - User Manager for UID 500... Nov 24 00:27:15.796317 (systemd)[1707]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Nov 24 00:27:15.799399 systemd-logind[1525]: New session c1 of user core. Nov 24 00:27:15.860009 systemd-timesyncd[1458]: Network configuration changed, trying to establish connection. Nov 24 00:27:15.936947 systemd[1707]: Queued start job for default target default.target. Nov 24 00:27:15.943896 systemd[1707]: Created slice app.slice - User Application Slice. Nov 24 00:27:15.943919 systemd[1707]: Reached target paths.target - Paths. Nov 24 00:27:15.944049 systemd[1707]: Reached target timers.target - Timers. Nov 24 00:27:15.945479 systemd[1707]: Starting dbus.socket - D-Bus User Message Bus Socket... Nov 24 00:27:15.957488 systemd[1707]: Listening on dbus.socket - D-Bus User Message Bus Socket. Nov 24 00:27:15.957611 systemd[1707]: Reached target sockets.target - Sockets. Nov 24 00:27:15.957928 systemd[1707]: Reached target basic.target - Basic System. Nov 24 00:27:15.957991 systemd[1707]: Reached target default.target - Main User Target. Nov 24 00:27:15.958024 systemd[1707]: Startup finished in 152ms. Nov 24 00:27:15.958107 systemd[1]: Started user@500.service - User Manager for UID 500. Nov 24 00:27:15.969829 systemd[1]: Started session-1.scope - Session 1 of User core. Nov 24 00:27:16.234812 systemd[1]: Started sshd@1-172.237.154.236:22-139.178.68.195:33520.service - OpenSSH per-connection server daemon (139.178.68.195:33520). Nov 24 00:27:16.575892 sshd[1718]: Accepted publickey for core from 139.178.68.195 port 33520 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:27:16.577696 sshd-session[1718]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:27:16.582588 systemd-logind[1525]: New session 2 of user core. Nov 24 00:27:16.588818 systemd[1]: Started session-2.scope - Session 2 of User core. Nov 24 00:27:16.829208 sshd[1721]: Connection closed by 139.178.68.195 port 33520 Nov 24 00:27:16.830135 sshd-session[1718]: pam_unix(sshd:session): session closed for user core Nov 24 00:27:16.833902 systemd[1]: sshd@1-172.237.154.236:22-139.178.68.195:33520.service: Deactivated successfully. Nov 24 00:27:16.835664 systemd[1]: session-2.scope: Deactivated successfully. Nov 24 00:27:16.836538 systemd-logind[1525]: Session 2 logged out. Waiting for processes to exit. Nov 24 00:27:16.838011 systemd-logind[1525]: Removed session 2. Nov 24 00:27:16.890973 systemd[1]: Started sshd@2-172.237.154.236:22-139.178.68.195:33530.service - OpenSSH per-connection server daemon (139.178.68.195:33530). Nov 24 00:27:17.244125 sshd[1727]: Accepted publickey for core from 139.178.68.195 port 33530 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:27:17.245901 sshd-session[1727]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:27:17.250908 systemd-logind[1525]: New session 3 of user core. Nov 24 00:27:17.260819 systemd[1]: Started session-3.scope - Session 3 of User core. Nov 24 00:27:17.495107 sshd[1730]: Connection closed by 139.178.68.195 port 33530 Nov 24 00:27:17.495902 sshd-session[1727]: pam_unix(sshd:session): session closed for user core Nov 24 00:27:17.499993 systemd[1]: sshd@2-172.237.154.236:22-139.178.68.195:33530.service: Deactivated successfully. Nov 24 00:27:17.500348 systemd-logind[1525]: Session 3 logged out. Waiting for processes to exit. Nov 24 00:27:17.502098 systemd[1]: session-3.scope: Deactivated successfully. Nov 24 00:27:17.503998 systemd-logind[1525]: Removed session 3. Nov 24 00:27:17.571538 systemd[1]: Started sshd@3-172.237.154.236:22-139.178.68.195:33544.service - OpenSSH per-connection server daemon (139.178.68.195:33544). Nov 24 00:27:17.932062 sshd[1736]: Accepted publickey for core from 139.178.68.195 port 33544 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:27:17.933507 sshd-session[1736]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:27:17.938905 systemd-logind[1525]: New session 4 of user core. Nov 24 00:27:17.943854 systemd[1]: Started session-4.scope - Session 4 of User core. Nov 24 00:27:18.192467 sshd[1739]: Connection closed by 139.178.68.195 port 33544 Nov 24 00:27:18.193163 sshd-session[1736]: pam_unix(sshd:session): session closed for user core Nov 24 00:27:18.197577 systemd-logind[1525]: Session 4 logged out. Waiting for processes to exit. Nov 24 00:27:18.198794 systemd[1]: sshd@3-172.237.154.236:22-139.178.68.195:33544.service: Deactivated successfully. Nov 24 00:27:18.200652 systemd[1]: session-4.scope: Deactivated successfully. Nov 24 00:27:18.201900 systemd-logind[1525]: Removed session 4. Nov 24 00:27:18.258807 systemd[1]: Started sshd@4-172.237.154.236:22-139.178.68.195:33546.service - OpenSSH per-connection server daemon (139.178.68.195:33546). Nov 24 00:27:18.605894 sshd[1745]: Accepted publickey for core from 139.178.68.195 port 33546 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:27:18.607998 sshd-session[1745]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:27:18.612923 systemd-logind[1525]: New session 5 of user core. Nov 24 00:27:18.619901 systemd[1]: Started session-5.scope - Session 5 of User core. Nov 24 00:27:18.816840 sudo[1749]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Nov 24 00:27:18.817160 sudo[1749]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 24 00:27:18.834115 sudo[1749]: pam_unix(sudo:session): session closed for user root Nov 24 00:27:18.887051 sshd[1748]: Connection closed by 139.178.68.195 port 33546 Nov 24 00:27:18.888212 sshd-session[1745]: pam_unix(sshd:session): session closed for user core Nov 24 00:27:18.898017 systemd[1]: sshd@4-172.237.154.236:22-139.178.68.195:33546.service: Deactivated successfully. Nov 24 00:27:18.900082 systemd[1]: session-5.scope: Deactivated successfully. Nov 24 00:27:18.901899 systemd-logind[1525]: Session 5 logged out. Waiting for processes to exit. Nov 24 00:27:18.903195 systemd-logind[1525]: Removed session 5. Nov 24 00:27:18.945125 systemd[1]: Started sshd@5-172.237.154.236:22-139.178.68.195:33562.service - OpenSSH per-connection server daemon (139.178.68.195:33562). Nov 24 00:27:19.285281 sshd[1755]: Accepted publickey for core from 139.178.68.195 port 33562 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:27:19.294432 sshd-session[1755]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:27:19.300926 systemd-logind[1525]: New session 6 of user core. Nov 24 00:27:19.306835 systemd[1]: Started session-6.scope - Session 6 of User core. Nov 24 00:27:19.483968 sudo[1760]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Nov 24 00:27:19.484289 sudo[1760]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 24 00:27:19.488900 sudo[1760]: pam_unix(sudo:session): session closed for user root Nov 24 00:27:19.494210 sudo[1759]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Nov 24 00:27:19.494513 sudo[1759]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 24 00:27:19.503911 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 24 00:27:19.541388 augenrules[1782]: No rules Nov 24 00:27:19.542419 systemd[1]: audit-rules.service: Deactivated successfully. Nov 24 00:27:19.542659 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 24 00:27:19.543419 sudo[1759]: pam_unix(sudo:session): session closed for user root Nov 24 00:27:19.593648 sshd[1758]: Connection closed by 139.178.68.195 port 33562 Nov 24 00:27:19.594103 sshd-session[1755]: pam_unix(sshd:session): session closed for user core Nov 24 00:27:19.598568 systemd-logind[1525]: Session 6 logged out. Waiting for processes to exit. Nov 24 00:27:19.598782 systemd[1]: sshd@5-172.237.154.236:22-139.178.68.195:33562.service: Deactivated successfully. Nov 24 00:27:19.600479 systemd[1]: session-6.scope: Deactivated successfully. Nov 24 00:27:19.602160 systemd-logind[1525]: Removed session 6. Nov 24 00:27:19.667985 systemd[1]: Started sshd@6-172.237.154.236:22-139.178.68.195:41572.service - OpenSSH per-connection server daemon (139.178.68.195:41572). Nov 24 00:27:20.024021 sshd[1791]: Accepted publickey for core from 139.178.68.195 port 41572 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:27:20.025250 sshd-session[1791]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:27:20.030606 systemd-logind[1525]: New session 7 of user core. Nov 24 00:27:20.037821 systemd[1]: Started session-7.scope - Session 7 of User core. Nov 24 00:27:20.232303 sudo[1795]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Nov 24 00:27:20.232626 sudo[1795]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 24 00:27:20.521252 systemd[1]: Starting docker.service - Docker Application Container Engine... Nov 24 00:27:20.532065 (dockerd)[1813]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Nov 24 00:27:20.747609 dockerd[1813]: time="2025-11-24T00:27:20.747556127Z" level=info msg="Starting up" Nov 24 00:27:20.749229 dockerd[1813]: time="2025-11-24T00:27:20.749211888Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Nov 24 00:27:20.761233 dockerd[1813]: time="2025-11-24T00:27:20.761197864Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Nov 24 00:27:20.776474 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2338207315-merged.mount: Deactivated successfully. Nov 24 00:27:20.804322 dockerd[1813]: time="2025-11-24T00:27:20.804268735Z" level=info msg="Loading containers: start." Nov 24 00:27:20.814847 kernel: Initializing XFRM netlink socket Nov 24 00:27:21.021205 systemd-timesyncd[1458]: Network configuration changed, trying to establish connection. Nov 24 00:27:21.022952 systemd-timesyncd[1458]: Network configuration changed, trying to establish connection. Nov 24 00:27:21.034194 systemd-timesyncd[1458]: Network configuration changed, trying to establish connection. Nov 24 00:27:21.075964 systemd-networkd[1439]: docker0: Link UP Nov 24 00:27:21.076319 systemd-timesyncd[1458]: Network configuration changed, trying to establish connection. Nov 24 00:27:21.078998 dockerd[1813]: time="2025-11-24T00:27:21.078943293Z" level=info msg="Loading containers: done." Nov 24 00:27:21.096348 dockerd[1813]: time="2025-11-24T00:27:21.096013741Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Nov 24 00:27:21.096348 dockerd[1813]: time="2025-11-24T00:27:21.096097301Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Nov 24 00:27:21.096348 dockerd[1813]: time="2025-11-24T00:27:21.096175891Z" level=info msg="Initializing buildkit" Nov 24 00:27:21.096978 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3012237694-merged.mount: Deactivated successfully. Nov 24 00:27:21.120414 dockerd[1813]: time="2025-11-24T00:27:21.120369553Z" level=info msg="Completed buildkit initialization" Nov 24 00:27:21.129129 dockerd[1813]: time="2025-11-24T00:27:21.129094778Z" level=info msg="Daemon has completed initialization" Nov 24 00:27:21.129279 dockerd[1813]: time="2025-11-24T00:27:21.129195388Z" level=info msg="API listen on /run/docker.sock" Nov 24 00:27:21.129398 systemd[1]: Started docker.service - Docker Application Container Engine. Nov 24 00:27:21.702483 containerd[1559]: time="2025-11-24T00:27:21.702395014Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.10\"" Nov 24 00:27:22.459086 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2465013239.mount: Deactivated successfully. Nov 24 00:27:23.639353 containerd[1559]: time="2025-11-24T00:27:23.639300572Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:23.640144 containerd[1559]: time="2025-11-24T00:27:23.640121812Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.10: active requests=0, bytes read=29072183" Nov 24 00:27:23.640793 containerd[1559]: time="2025-11-24T00:27:23.640768913Z" level=info msg="ImageCreate event name:\"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:23.642804 containerd[1559]: time="2025-11-24T00:27:23.642782874Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:af4ee57c047e31a7f58422b94a9ec4c62221d3deebb16755bdeff720df796189\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:23.643600 containerd[1559]: time="2025-11-24T00:27:23.643573134Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.10\" with image id \"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:af4ee57c047e31a7f58422b94a9ec4c62221d3deebb16755bdeff720df796189\", size \"29068782\" in 1.94113145s" Nov 24 00:27:23.643643 containerd[1559]: time="2025-11-24T00:27:23.643606334Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.10\" returns image reference \"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\"" Nov 24 00:27:23.644729 containerd[1559]: time="2025-11-24T00:27:23.644690065Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.10\"" Nov 24 00:27:25.007115 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Nov 24 00:27:25.010643 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:27:25.217845 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:27:25.227210 (kubelet)[2094]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 24 00:27:25.275792 containerd[1559]: time="2025-11-24T00:27:25.274648069Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:25.276610 containerd[1559]: time="2025-11-24T00:27:25.276593360Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.10: active requests=0, bytes read=24992010" Nov 24 00:27:25.277293 containerd[1559]: time="2025-11-24T00:27:25.277275110Z" level=info msg="ImageCreate event name:\"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:25.280124 containerd[1559]: time="2025-11-24T00:27:25.280104902Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:efbd9d1dfcd2940e1c73a1476c880c3c2cdf04cc60722d329b21cd48745c8660\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:25.281652 containerd[1559]: time="2025-11-24T00:27:25.281621462Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.10\" with image id \"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:efbd9d1dfcd2940e1c73a1476c880c3c2cdf04cc60722d329b21cd48745c8660\", size \"26649046\" in 1.636852427s" Nov 24 00:27:25.281946 containerd[1559]: time="2025-11-24T00:27:25.281927723Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.10\" returns image reference \"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\"" Nov 24 00:27:25.282809 containerd[1559]: time="2025-11-24T00:27:25.282773443Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.10\"" Nov 24 00:27:25.288683 kubelet[2094]: E1124 00:27:25.288630 2094 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 24 00:27:25.293792 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 24 00:27:25.293993 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 24 00:27:25.294482 systemd[1]: kubelet.service: Consumed 205ms CPU time, 108.5M memory peak. Nov 24 00:27:26.483022 containerd[1559]: time="2025-11-24T00:27:26.482944303Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:26.483902 containerd[1559]: time="2025-11-24T00:27:26.483882273Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.10: active requests=0, bytes read=19404248" Nov 24 00:27:26.484418 containerd[1559]: time="2025-11-24T00:27:26.484383143Z" level=info msg="ImageCreate event name:\"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:26.486852 containerd[1559]: time="2025-11-24T00:27:26.486614544Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:9c58e1adcad5af66d1d9ca5cf9a4c266e4054b8f19f91a8fff1993549e657b10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:26.488098 containerd[1559]: time="2025-11-24T00:27:26.487791015Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.10\" with image id \"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:9c58e1adcad5af66d1d9ca5cf9a4c266e4054b8f19f91a8fff1993549e657b10\", size \"21061302\" in 1.204888922s" Nov 24 00:27:26.488098 containerd[1559]: time="2025-11-24T00:27:26.487816745Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.10\" returns image reference \"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\"" Nov 24 00:27:26.488267 containerd[1559]: time="2025-11-24T00:27:26.488219825Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.10\"" Nov 24 00:27:27.581496 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2953140250.mount: Deactivated successfully. Nov 24 00:27:27.904980 containerd[1559]: time="2025-11-24T00:27:27.904399123Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:27.905387 containerd[1559]: time="2025-11-24T00:27:27.905318163Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.10: active requests=0, bytes read=31161423" Nov 24 00:27:27.905923 containerd[1559]: time="2025-11-24T00:27:27.905893424Z" level=info msg="ImageCreate event name:\"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:27.907343 containerd[1559]: time="2025-11-24T00:27:27.907323854Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e3dda1c7b384f9eb5b2fa1c27493b23b80e6204b9fa2ee8791b2de078f468cbf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:27.908095 containerd[1559]: time="2025-11-24T00:27:27.908060865Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.10\" with image id \"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\", repo tag \"registry.k8s.io/kube-proxy:v1.32.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:e3dda1c7b384f9eb5b2fa1c27493b23b80e6204b9fa2ee8791b2de078f468cbf\", size \"31160442\" in 1.41962143s" Nov 24 00:27:27.908131 containerd[1559]: time="2025-11-24T00:27:27.908094665Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.10\" returns image reference \"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\"" Nov 24 00:27:27.908550 containerd[1559]: time="2025-11-24T00:27:27.908465115Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Nov 24 00:27:28.542413 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3598038495.mount: Deactivated successfully. Nov 24 00:27:29.238035 containerd[1559]: time="2025-11-24T00:27:29.237185189Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:29.238035 containerd[1559]: time="2025-11-24T00:27:29.237986219Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Nov 24 00:27:29.238866 containerd[1559]: time="2025-11-24T00:27:29.238842160Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:29.243153 containerd[1559]: time="2025-11-24T00:27:29.243124632Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:29.244836 containerd[1559]: time="2025-11-24T00:27:29.244811023Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.336315918s" Nov 24 00:27:29.244923 containerd[1559]: time="2025-11-24T00:27:29.244907503Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Nov 24 00:27:29.245413 containerd[1559]: time="2025-11-24T00:27:29.245387443Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Nov 24 00:27:29.840465 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount517384616.mount: Deactivated successfully. Nov 24 00:27:29.845742 containerd[1559]: time="2025-11-24T00:27:29.845081113Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 24 00:27:29.845880 containerd[1559]: time="2025-11-24T00:27:29.845836553Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Nov 24 00:27:29.846417 containerd[1559]: time="2025-11-24T00:27:29.846396023Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 24 00:27:29.848067 containerd[1559]: time="2025-11-24T00:27:29.848046744Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 24 00:27:29.848804 containerd[1559]: time="2025-11-24T00:27:29.848481064Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 602.963271ms" Nov 24 00:27:29.848804 containerd[1559]: time="2025-11-24T00:27:29.848511074Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Nov 24 00:27:29.848997 containerd[1559]: time="2025-11-24T00:27:29.848973444Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Nov 24 00:27:30.549320 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount484086723.mount: Deactivated successfully. Nov 24 00:27:32.019495 containerd[1559]: time="2025-11-24T00:27:32.019412109Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:32.020576 containerd[1559]: time="2025-11-24T00:27:32.020543150Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Nov 24 00:27:32.021274 containerd[1559]: time="2025-11-24T00:27:32.021233280Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:32.023786 containerd[1559]: time="2025-11-24T00:27:32.023746271Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:32.024788 containerd[1559]: time="2025-11-24T00:27:32.024601982Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.175530977s" Nov 24 00:27:32.024788 containerd[1559]: time="2025-11-24T00:27:32.024641732Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Nov 24 00:27:33.895627 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:27:33.896107 systemd[1]: kubelet.service: Consumed 205ms CPU time, 108.5M memory peak. Nov 24 00:27:33.899247 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:27:33.927300 systemd[1]: Reload requested from client PID 2251 ('systemctl') (unit session-7.scope)... Nov 24 00:27:33.927316 systemd[1]: Reloading... Nov 24 00:27:34.056927 zram_generator::config[2292]: No configuration found. Nov 24 00:27:34.276955 systemd[1]: Reloading finished in 349 ms. Nov 24 00:27:34.334225 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Nov 24 00:27:34.334328 systemd[1]: kubelet.service: Failed with result 'signal'. Nov 24 00:27:34.334647 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:27:34.334688 systemd[1]: kubelet.service: Consumed 133ms CPU time, 98.3M memory peak. Nov 24 00:27:34.336376 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:27:34.506189 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:27:34.515178 (kubelet)[2349]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 24 00:27:34.550695 kubelet[2349]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:27:34.552733 kubelet[2349]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 24 00:27:34.552733 kubelet[2349]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:27:34.552733 kubelet[2349]: I1124 00:27:34.551041 2349 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 00:27:34.830614 kubelet[2349]: I1124 00:27:34.830484 2349 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 24 00:27:34.830741 kubelet[2349]: I1124 00:27:34.830728 2349 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 00:27:34.830986 kubelet[2349]: I1124 00:27:34.830975 2349 server.go:954] "Client rotation is on, will bootstrap in background" Nov 24 00:27:34.858852 kubelet[2349]: E1124 00:27:34.858825 2349 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.237.154.236:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.237.154.236:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:27:34.859165 kubelet[2349]: I1124 00:27:34.859145 2349 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 24 00:27:34.865038 kubelet[2349]: I1124 00:27:34.865021 2349 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 00:27:34.872117 kubelet[2349]: I1124 00:27:34.872073 2349 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 24 00:27:34.872526 kubelet[2349]: I1124 00:27:34.872503 2349 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 00:27:34.872662 kubelet[2349]: I1124 00:27:34.872525 2349 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"172-237-154-236","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 00:27:34.872788 kubelet[2349]: I1124 00:27:34.872667 2349 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 00:27:34.872788 kubelet[2349]: I1124 00:27:34.872675 2349 container_manager_linux.go:304] "Creating device plugin manager" Nov 24 00:27:34.872832 kubelet[2349]: I1124 00:27:34.872808 2349 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:27:34.876368 kubelet[2349]: I1124 00:27:34.876256 2349 kubelet.go:446] "Attempting to sync node with API server" Nov 24 00:27:34.878383 kubelet[2349]: I1124 00:27:34.878044 2349 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 00:27:34.878383 kubelet[2349]: I1124 00:27:34.878073 2349 kubelet.go:352] "Adding apiserver pod source" Nov 24 00:27:34.878383 kubelet[2349]: I1124 00:27:34.878084 2349 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 00:27:34.879590 kubelet[2349]: W1124 00:27:34.879552 2349 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.237.154.236:6443/api/v1/nodes?fieldSelector=metadata.name%3D172-237-154-236&limit=500&resourceVersion=0": dial tcp 172.237.154.236:6443: connect: connection refused Nov 24 00:27:34.879672 kubelet[2349]: E1124 00:27:34.879657 2349 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.237.154.236:6443/api/v1/nodes?fieldSelector=metadata.name%3D172-237-154-236&limit=500&resourceVersion=0\": dial tcp 172.237.154.236:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:27:34.880388 kubelet[2349]: W1124 00:27:34.880356 2349 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.237.154.236:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.237.154.236:6443: connect: connection refused Nov 24 00:27:34.880440 kubelet[2349]: E1124 00:27:34.880392 2349 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.237.154.236:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.237.154.236:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:27:34.880700 kubelet[2349]: I1124 00:27:34.880680 2349 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Nov 24 00:27:34.881169 kubelet[2349]: I1124 00:27:34.881153 2349 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 00:27:34.881915 kubelet[2349]: W1124 00:27:34.881899 2349 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Nov 24 00:27:34.885151 kubelet[2349]: I1124 00:27:34.884591 2349 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 24 00:27:34.885151 kubelet[2349]: I1124 00:27:34.884620 2349 server.go:1287] "Started kubelet" Nov 24 00:27:34.888033 kubelet[2349]: I1124 00:27:34.888008 2349 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 00:27:34.888829 kubelet[2349]: I1124 00:27:34.888787 2349 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 00:27:34.888888 kubelet[2349]: I1124 00:27:34.888810 2349 server.go:479] "Adding debug handlers to kubelet server" Nov 24 00:27:34.889107 kubelet[2349]: I1124 00:27:34.889088 2349 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 00:27:34.890434 kubelet[2349]: E1124 00:27:34.889223 2349 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.237.154.236:6443/api/v1/namespaces/default/events\": dial tcp 172.237.154.236:6443: connect: connection refused" event="&Event{ObjectMeta:{172-237-154-236.187ac9c5ab0f1f73 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172-237-154-236,UID:172-237-154-236,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:172-237-154-236,},FirstTimestamp:2025-11-24 00:27:34.884605811 +0000 UTC m=+0.365499564,LastTimestamp:2025-11-24 00:27:34.884605811 +0000 UTC m=+0.365499564,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172-237-154-236,}" Nov 24 00:27:34.891081 kubelet[2349]: I1124 00:27:34.891068 2349 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 00:27:34.892255 kubelet[2349]: I1124 00:27:34.892240 2349 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 24 00:27:34.895894 kubelet[2349]: E1124 00:27:34.895089 2349 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172-237-154-236\" not found" Nov 24 00:27:34.895894 kubelet[2349]: I1124 00:27:34.895112 2349 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 24 00:27:34.895894 kubelet[2349]: I1124 00:27:34.895211 2349 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 24 00:27:34.895894 kubelet[2349]: I1124 00:27:34.895246 2349 reconciler.go:26] "Reconciler: start to sync state" Nov 24 00:27:34.895894 kubelet[2349]: W1124 00:27:34.895451 2349 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.237.154.236:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.237.154.236:6443: connect: connection refused Nov 24 00:27:34.895894 kubelet[2349]: E1124 00:27:34.895477 2349 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.237.154.236:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.237.154.236:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:27:34.896230 kubelet[2349]: E1124 00:27:34.896209 2349 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.237.154.236:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-237-154-236?timeout=10s\": dial tcp 172.237.154.236:6443: connect: connection refused" interval="200ms" Nov 24 00:27:34.896764 kubelet[2349]: I1124 00:27:34.896747 2349 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 24 00:27:34.897138 kubelet[2349]: E1124 00:27:34.897124 2349 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 24 00:27:34.898465 kubelet[2349]: I1124 00:27:34.898451 2349 factory.go:221] Registration of the containerd container factory successfully Nov 24 00:27:34.898549 kubelet[2349]: I1124 00:27:34.898540 2349 factory.go:221] Registration of the systemd container factory successfully Nov 24 00:27:34.912495 kubelet[2349]: I1124 00:27:34.912469 2349 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 00:27:34.913971 kubelet[2349]: I1124 00:27:34.913956 2349 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 00:27:34.914029 kubelet[2349]: I1124 00:27:34.914020 2349 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 24 00:27:34.914078 kubelet[2349]: I1124 00:27:34.914069 2349 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 24 00:27:34.914134 kubelet[2349]: I1124 00:27:34.914124 2349 kubelet.go:2382] "Starting kubelet main sync loop" Nov 24 00:27:34.914219 kubelet[2349]: E1124 00:27:34.914202 2349 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 00:27:34.924965 kubelet[2349]: W1124 00:27:34.922832 2349 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.237.154.236:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.237.154.236:6443: connect: connection refused Nov 24 00:27:34.924965 kubelet[2349]: E1124 00:27:34.924961 2349 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.237.154.236:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.237.154.236:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:27:34.933131 kubelet[2349]: I1124 00:27:34.933103 2349 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 24 00:27:34.933272 kubelet[2349]: I1124 00:27:34.933261 2349 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 24 00:27:34.933350 kubelet[2349]: I1124 00:27:34.933340 2349 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:27:34.935007 kubelet[2349]: I1124 00:27:34.934994 2349 policy_none.go:49] "None policy: Start" Nov 24 00:27:34.935095 kubelet[2349]: I1124 00:27:34.935086 2349 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 24 00:27:34.935236 kubelet[2349]: I1124 00:27:34.935138 2349 state_mem.go:35] "Initializing new in-memory state store" Nov 24 00:27:34.941676 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Nov 24 00:27:34.951524 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Nov 24 00:27:34.972798 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Nov 24 00:27:34.974909 kubelet[2349]: I1124 00:27:34.974637 2349 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 00:27:34.975056 kubelet[2349]: I1124 00:27:34.975027 2349 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 00:27:34.975095 kubelet[2349]: I1124 00:27:34.975045 2349 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 00:27:34.975353 kubelet[2349]: I1124 00:27:34.975334 2349 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 00:27:34.976111 kubelet[2349]: E1124 00:27:34.976097 2349 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 24 00:27:34.976490 kubelet[2349]: E1124 00:27:34.976209 2349 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"172-237-154-236\" not found" Nov 24 00:27:35.024281 systemd[1]: Created slice kubepods-burstable-poddcd4b3957f17b017ab6167b0ac63a70f.slice - libcontainer container kubepods-burstable-poddcd4b3957f17b017ab6167b0ac63a70f.slice. Nov 24 00:27:35.045379 kubelet[2349]: E1124 00:27:35.045352 2349 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:35.048435 systemd[1]: Created slice kubepods-burstable-pod973cb66da29b95ef423bd337cf05537d.slice - libcontainer container kubepods-burstable-pod973cb66da29b95ef423bd337cf05537d.slice. Nov 24 00:27:35.057409 kubelet[2349]: E1124 00:27:35.057246 2349 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:35.060296 systemd[1]: Created slice kubepods-burstable-pod6d5965ce9dc5a2c239f1085a4e76621b.slice - libcontainer container kubepods-burstable-pod6d5965ce9dc5a2c239f1085a4e76621b.slice. Nov 24 00:27:35.061960 kubelet[2349]: E1124 00:27:35.061947 2349 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:35.077271 kubelet[2349]: I1124 00:27:35.077257 2349 kubelet_node_status.go:75] "Attempting to register node" node="172-237-154-236" Nov 24 00:27:35.077511 kubelet[2349]: E1124 00:27:35.077490 2349 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.237.154.236:6443/api/v1/nodes\": dial tcp 172.237.154.236:6443: connect: connection refused" node="172-237-154-236" Nov 24 00:27:35.097056 kubelet[2349]: E1124 00:27:35.096972 2349 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.237.154.236:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-237-154-236?timeout=10s\": dial tcp 172.237.154.236:6443: connect: connection refused" interval="400ms" Nov 24 00:27:35.196365 kubelet[2349]: I1124 00:27:35.196286 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-ca-certs\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:35.196365 kubelet[2349]: I1124 00:27:35.196314 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-k8s-certs\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:35.196365 kubelet[2349]: I1124 00:27:35.196351 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-kubeconfig\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:35.196478 kubelet[2349]: I1124 00:27:35.196368 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-usr-share-ca-certificates\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:35.196478 kubelet[2349]: I1124 00:27:35.196393 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6d5965ce9dc5a2c239f1085a4e76621b-kubeconfig\") pod \"kube-scheduler-172-237-154-236\" (UID: \"6d5965ce9dc5a2c239f1085a4e76621b\") " pod="kube-system/kube-scheduler-172-237-154-236" Nov 24 00:27:35.196478 kubelet[2349]: I1124 00:27:35.196406 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/dcd4b3957f17b017ab6167b0ac63a70f-k8s-certs\") pod \"kube-apiserver-172-237-154-236\" (UID: \"dcd4b3957f17b017ab6167b0ac63a70f\") " pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:35.196478 kubelet[2349]: I1124 00:27:35.196441 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/dcd4b3957f17b017ab6167b0ac63a70f-usr-share-ca-certificates\") pod \"kube-apiserver-172-237-154-236\" (UID: \"dcd4b3957f17b017ab6167b0ac63a70f\") " pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:35.196478 kubelet[2349]: I1124 00:27:35.196454 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/dcd4b3957f17b017ab6167b0ac63a70f-ca-certs\") pod \"kube-apiserver-172-237-154-236\" (UID: \"dcd4b3957f17b017ab6167b0ac63a70f\") " pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:35.196600 kubelet[2349]: I1124 00:27:35.196468 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-flexvolume-dir\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:35.280015 kubelet[2349]: I1124 00:27:35.279851 2349 kubelet_node_status.go:75] "Attempting to register node" node="172-237-154-236" Nov 24 00:27:35.280094 kubelet[2349]: E1124 00:27:35.280071 2349 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.237.154.236:6443/api/v1/nodes\": dial tcp 172.237.154.236:6443: connect: connection refused" node="172-237-154-236" Nov 24 00:27:35.346676 kubelet[2349]: E1124 00:27:35.346645 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:35.347161 containerd[1559]: time="2025-11-24T00:27:35.347064702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-172-237-154-236,Uid:dcd4b3957f17b017ab6167b0ac63a70f,Namespace:kube-system,Attempt:0,}" Nov 24 00:27:35.360348 kubelet[2349]: E1124 00:27:35.360283 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:35.361683 containerd[1559]: time="2025-11-24T00:27:35.360859529Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-172-237-154-236,Uid:973cb66da29b95ef423bd337cf05537d,Namespace:kube-system,Attempt:0,}" Nov 24 00:27:35.363134 kubelet[2349]: E1124 00:27:35.363093 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:35.363812 containerd[1559]: time="2025-11-24T00:27:35.363752770Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-172-237-154-236,Uid:6d5965ce9dc5a2c239f1085a4e76621b,Namespace:kube-system,Attempt:0,}" Nov 24 00:27:35.370272 containerd[1559]: time="2025-11-24T00:27:35.370230173Z" level=info msg="connecting to shim 57ea78cdf24f6e0d97684a0af0fbdebc84fd292d19ae8edcb3aefea384aeaea4" address="unix:///run/containerd/s/66eeadab593349df837aa7c161eaeab36b78116590a4a9b718b225ea759727bd" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:27:35.404508 containerd[1559]: time="2025-11-24T00:27:35.404464490Z" level=info msg="connecting to shim 1728c0f60814fa7dd9b3d261cb1e30ea90c7e028648c2c3f4e6a9853c457eff4" address="unix:///run/containerd/s/bc3af573ccdab649074b3c3230668efbd0e5df6ae0f1f160bea86eb097be7ef0" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:27:35.411630 containerd[1559]: time="2025-11-24T00:27:35.411550034Z" level=info msg="connecting to shim 2421d147d5b524c062137c6d0d166d06476b1ce823fe1e48d088407231a483ad" address="unix:///run/containerd/s/c8d6d72555d45aaf6a5b01ba2a520ecf4f0fad457c66118f6b9d10bd820bfe9e" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:27:35.423851 systemd[1]: Started cri-containerd-57ea78cdf24f6e0d97684a0af0fbdebc84fd292d19ae8edcb3aefea384aeaea4.scope - libcontainer container 57ea78cdf24f6e0d97684a0af0fbdebc84fd292d19ae8edcb3aefea384aeaea4. Nov 24 00:27:35.441199 systemd[1]: Started cri-containerd-1728c0f60814fa7dd9b3d261cb1e30ea90c7e028648c2c3f4e6a9853c457eff4.scope - libcontainer container 1728c0f60814fa7dd9b3d261cb1e30ea90c7e028648c2c3f4e6a9853c457eff4. Nov 24 00:27:35.458954 systemd[1]: Started cri-containerd-2421d147d5b524c062137c6d0d166d06476b1ce823fe1e48d088407231a483ad.scope - libcontainer container 2421d147d5b524c062137c6d0d166d06476b1ce823fe1e48d088407231a483ad. Nov 24 00:27:35.500348 kubelet[2349]: E1124 00:27:35.500284 2349 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.237.154.236:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-237-154-236?timeout=10s\": dial tcp 172.237.154.236:6443: connect: connection refused" interval="800ms" Nov 24 00:27:35.532386 containerd[1559]: time="2025-11-24T00:27:35.532345924Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-172-237-154-236,Uid:dcd4b3957f17b017ab6167b0ac63a70f,Namespace:kube-system,Attempt:0,} returns sandbox id \"57ea78cdf24f6e0d97684a0af0fbdebc84fd292d19ae8edcb3aefea384aeaea4\"" Nov 24 00:27:35.535467 kubelet[2349]: E1124 00:27:35.535424 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:35.537188 containerd[1559]: time="2025-11-24T00:27:35.537159777Z" level=info msg="CreateContainer within sandbox \"57ea78cdf24f6e0d97684a0af0fbdebc84fd292d19ae8edcb3aefea384aeaea4\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Nov 24 00:27:35.545168 containerd[1559]: time="2025-11-24T00:27:35.545090011Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-172-237-154-236,Uid:973cb66da29b95ef423bd337cf05537d,Namespace:kube-system,Attempt:0,} returns sandbox id \"1728c0f60814fa7dd9b3d261cb1e30ea90c7e028648c2c3f4e6a9853c457eff4\"" Nov 24 00:27:35.546043 kubelet[2349]: E1124 00:27:35.545945 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:35.547549 containerd[1559]: time="2025-11-24T00:27:35.547362982Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-172-237-154-236,Uid:6d5965ce9dc5a2c239f1085a4e76621b,Namespace:kube-system,Attempt:0,} returns sandbox id \"2421d147d5b524c062137c6d0d166d06476b1ce823fe1e48d088407231a483ad\"" Nov 24 00:27:35.548697 kubelet[2349]: E1124 00:27:35.548683 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:35.550456 containerd[1559]: time="2025-11-24T00:27:35.549587583Z" level=info msg="Container 919e0cbaae6c04972adec8f5617be183d926cf2289c28dd49bd7ba95007ac8b3: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:27:35.550637 containerd[1559]: time="2025-11-24T00:27:35.550617653Z" level=info msg="CreateContainer within sandbox \"1728c0f60814fa7dd9b3d261cb1e30ea90c7e028648c2c3f4e6a9853c457eff4\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Nov 24 00:27:35.555107 containerd[1559]: time="2025-11-24T00:27:35.555085816Z" level=info msg="CreateContainer within sandbox \"2421d147d5b524c062137c6d0d166d06476b1ce823fe1e48d088407231a483ad\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Nov 24 00:27:35.559491 containerd[1559]: time="2025-11-24T00:27:35.559470768Z" level=info msg="CreateContainer within sandbox \"57ea78cdf24f6e0d97684a0af0fbdebc84fd292d19ae8edcb3aefea384aeaea4\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"919e0cbaae6c04972adec8f5617be183d926cf2289c28dd49bd7ba95007ac8b3\"" Nov 24 00:27:35.560384 containerd[1559]: time="2025-11-24T00:27:35.560365788Z" level=info msg="StartContainer for \"919e0cbaae6c04972adec8f5617be183d926cf2289c28dd49bd7ba95007ac8b3\"" Nov 24 00:27:35.562741 containerd[1559]: time="2025-11-24T00:27:35.562684989Z" level=info msg="connecting to shim 919e0cbaae6c04972adec8f5617be183d926cf2289c28dd49bd7ba95007ac8b3" address="unix:///run/containerd/s/66eeadab593349df837aa7c161eaeab36b78116590a4a9b718b225ea759727bd" protocol=ttrpc version=3 Nov 24 00:27:35.565946 containerd[1559]: time="2025-11-24T00:27:35.565926751Z" level=info msg="Container b4f59269979b7792c20797631eae8255084097fc26d9e0824a0582b26353f44a: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:27:35.568648 containerd[1559]: time="2025-11-24T00:27:35.568625082Z" level=info msg="Container 56afb6e09a70a86fd5bbdfc83390eaa88b86adb754c08436977ec1d1b1d6e9e8: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:27:35.569883 containerd[1559]: time="2025-11-24T00:27:35.569863553Z" level=info msg="CreateContainer within sandbox \"1728c0f60814fa7dd9b3d261cb1e30ea90c7e028648c2c3f4e6a9853c457eff4\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b4f59269979b7792c20797631eae8255084097fc26d9e0824a0582b26353f44a\"" Nov 24 00:27:35.570339 containerd[1559]: time="2025-11-24T00:27:35.570322783Z" level=info msg="StartContainer for \"b4f59269979b7792c20797631eae8255084097fc26d9e0824a0582b26353f44a\"" Nov 24 00:27:35.571935 containerd[1559]: time="2025-11-24T00:27:35.571803804Z" level=info msg="connecting to shim b4f59269979b7792c20797631eae8255084097fc26d9e0824a0582b26353f44a" address="unix:///run/containerd/s/bc3af573ccdab649074b3c3230668efbd0e5df6ae0f1f160bea86eb097be7ef0" protocol=ttrpc version=3 Nov 24 00:27:35.577450 containerd[1559]: time="2025-11-24T00:27:35.577418767Z" level=info msg="CreateContainer within sandbox \"2421d147d5b524c062137c6d0d166d06476b1ce823fe1e48d088407231a483ad\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"56afb6e09a70a86fd5bbdfc83390eaa88b86adb754c08436977ec1d1b1d6e9e8\"" Nov 24 00:27:35.578462 containerd[1559]: time="2025-11-24T00:27:35.578431927Z" level=info msg="StartContainer for \"56afb6e09a70a86fd5bbdfc83390eaa88b86adb754c08436977ec1d1b1d6e9e8\"" Nov 24 00:27:35.580043 containerd[1559]: time="2025-11-24T00:27:35.580010388Z" level=info msg="connecting to shim 56afb6e09a70a86fd5bbdfc83390eaa88b86adb754c08436977ec1d1b1d6e9e8" address="unix:///run/containerd/s/c8d6d72555d45aaf6a5b01ba2a520ecf4f0fad457c66118f6b9d10bd820bfe9e" protocol=ttrpc version=3 Nov 24 00:27:35.591962 systemd[1]: Started cri-containerd-919e0cbaae6c04972adec8f5617be183d926cf2289c28dd49bd7ba95007ac8b3.scope - libcontainer container 919e0cbaae6c04972adec8f5617be183d926cf2289c28dd49bd7ba95007ac8b3. Nov 24 00:27:35.614087 systemd[1]: Started cri-containerd-56afb6e09a70a86fd5bbdfc83390eaa88b86adb754c08436977ec1d1b1d6e9e8.scope - libcontainer container 56afb6e09a70a86fd5bbdfc83390eaa88b86adb754c08436977ec1d1b1d6e9e8. Nov 24 00:27:35.618998 systemd[1]: Started cri-containerd-b4f59269979b7792c20797631eae8255084097fc26d9e0824a0582b26353f44a.scope - libcontainer container b4f59269979b7792c20797631eae8255084097fc26d9e0824a0582b26353f44a. Nov 24 00:27:35.681998 kubelet[2349]: I1124 00:27:35.681963 2349 kubelet_node_status.go:75] "Attempting to register node" node="172-237-154-236" Nov 24 00:27:35.682526 kubelet[2349]: E1124 00:27:35.682457 2349 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.237.154.236:6443/api/v1/nodes\": dial tcp 172.237.154.236:6443: connect: connection refused" node="172-237-154-236" Nov 24 00:27:35.684604 containerd[1559]: time="2025-11-24T00:27:35.684575080Z" level=info msg="StartContainer for \"919e0cbaae6c04972adec8f5617be183d926cf2289c28dd49bd7ba95007ac8b3\" returns successfully" Nov 24 00:27:35.702236 containerd[1559]: time="2025-11-24T00:27:35.702195569Z" level=info msg="StartContainer for \"b4f59269979b7792c20797631eae8255084097fc26d9e0824a0582b26353f44a\" returns successfully" Nov 24 00:27:35.723679 containerd[1559]: time="2025-11-24T00:27:35.723644740Z" level=info msg="StartContainer for \"56afb6e09a70a86fd5bbdfc83390eaa88b86adb754c08436977ec1d1b1d6e9e8\" returns successfully" Nov 24 00:27:35.938376 kubelet[2349]: E1124 00:27:35.938281 2349 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:35.938466 kubelet[2349]: E1124 00:27:35.938401 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:35.946152 kubelet[2349]: E1124 00:27:35.946133 2349 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:35.946230 kubelet[2349]: E1124 00:27:35.946213 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:35.946809 kubelet[2349]: E1124 00:27:35.946794 2349 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:35.946887 kubelet[2349]: E1124 00:27:35.946872 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:36.485307 kubelet[2349]: I1124 00:27:36.485279 2349 kubelet_node_status.go:75] "Attempting to register node" node="172-237-154-236" Nov 24 00:27:36.949637 kubelet[2349]: E1124 00:27:36.949459 2349 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:36.949637 kubelet[2349]: E1124 00:27:36.949579 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:36.950989 kubelet[2349]: E1124 00:27:36.950677 2349 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:36.951235 kubelet[2349]: E1124 00:27:36.951171 2349 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:37.374167 kubelet[2349]: E1124 00:27:37.373925 2349 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"172-237-154-236\" not found" node="172-237-154-236" Nov 24 00:27:37.436731 kubelet[2349]: I1124 00:27:37.436689 2349 kubelet_node_status.go:78] "Successfully registered node" node="172-237-154-236" Nov 24 00:27:37.496143 kubelet[2349]: I1124 00:27:37.496082 2349 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:37.551415 kubelet[2349]: E1124 00:27:37.551365 2349 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-172-237-154-236\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:37.551415 kubelet[2349]: I1124 00:27:37.551388 2349 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:37.553808 kubelet[2349]: E1124 00:27:37.553368 2349 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-172-237-154-236\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:37.553808 kubelet[2349]: I1124 00:27:37.553381 2349 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-172-237-154-236" Nov 24 00:27:37.559140 kubelet[2349]: E1124 00:27:37.559111 2349 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-172-237-154-236\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-172-237-154-236" Nov 24 00:27:37.881780 kubelet[2349]: I1124 00:27:37.881755 2349 apiserver.go:52] "Watching apiserver" Nov 24 00:27:37.896047 kubelet[2349]: I1124 00:27:37.896017 2349 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 24 00:27:39.456770 systemd[1]: Reload requested from client PID 2614 ('systemctl') (unit session-7.scope)... Nov 24 00:27:39.456789 systemd[1]: Reloading... Nov 24 00:27:39.568974 zram_generator::config[2667]: No configuration found. Nov 24 00:27:39.791829 systemd[1]: Reloading finished in 334 ms. Nov 24 00:27:39.820093 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:27:39.842804 systemd[1]: kubelet.service: Deactivated successfully. Nov 24 00:27:39.843084 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:27:39.843139 systemd[1]: kubelet.service: Consumed 733ms CPU time, 128.9M memory peak. Nov 24 00:27:39.845070 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:27:40.041102 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:27:40.050090 (kubelet)[2709]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 24 00:27:40.095270 kubelet[2709]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:27:40.095270 kubelet[2709]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 24 00:27:40.095270 kubelet[2709]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:27:40.095270 kubelet[2709]: I1124 00:27:40.095184 2709 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 00:27:40.103311 kubelet[2709]: I1124 00:27:40.103284 2709 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 24 00:27:40.103311 kubelet[2709]: I1124 00:27:40.103304 2709 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 00:27:40.103517 kubelet[2709]: I1124 00:27:40.103495 2709 server.go:954] "Client rotation is on, will bootstrap in background" Nov 24 00:27:40.104949 kubelet[2709]: I1124 00:27:40.104758 2709 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 00:27:40.107741 kubelet[2709]: I1124 00:27:40.107229 2709 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 24 00:27:40.112175 kubelet[2709]: I1124 00:27:40.112159 2709 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 00:27:40.119969 kubelet[2709]: I1124 00:27:40.119945 2709 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 24 00:27:40.120317 kubelet[2709]: I1124 00:27:40.120285 2709 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 00:27:40.120466 kubelet[2709]: I1124 00:27:40.120316 2709 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"172-237-154-236","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 00:27:40.120466 kubelet[2709]: I1124 00:27:40.120472 2709 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 00:27:40.120574 kubelet[2709]: I1124 00:27:40.120481 2709 container_manager_linux.go:304] "Creating device plugin manager" Nov 24 00:27:40.120574 kubelet[2709]: I1124 00:27:40.120552 2709 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:27:40.121377 kubelet[2709]: I1124 00:27:40.120765 2709 kubelet.go:446] "Attempting to sync node with API server" Nov 24 00:27:40.121377 kubelet[2709]: I1124 00:27:40.120788 2709 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 00:27:40.121377 kubelet[2709]: I1124 00:27:40.120807 2709 kubelet.go:352] "Adding apiserver pod source" Nov 24 00:27:40.121377 kubelet[2709]: I1124 00:27:40.120815 2709 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 00:27:40.125824 kubelet[2709]: I1124 00:27:40.125809 2709 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Nov 24 00:27:40.126315 kubelet[2709]: I1124 00:27:40.126298 2709 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 00:27:40.126781 kubelet[2709]: I1124 00:27:40.126768 2709 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 24 00:27:40.127058 kubelet[2709]: I1124 00:27:40.127044 2709 server.go:1287] "Started kubelet" Nov 24 00:27:40.127321 kubelet[2709]: I1124 00:27:40.127290 2709 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 00:27:40.127457 kubelet[2709]: I1124 00:27:40.127425 2709 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 00:27:40.127748 kubelet[2709]: I1124 00:27:40.127734 2709 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 00:27:40.128330 kubelet[2709]: I1124 00:27:40.128307 2709 server.go:479] "Adding debug handlers to kubelet server" Nov 24 00:27:40.130135 kubelet[2709]: I1124 00:27:40.130121 2709 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 00:27:40.133214 kubelet[2709]: I1124 00:27:40.132256 2709 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 24 00:27:40.137448 kubelet[2709]: I1124 00:27:40.137427 2709 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 24 00:27:40.137633 kubelet[2709]: I1124 00:27:40.137613 2709 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 24 00:27:40.138316 kubelet[2709]: I1124 00:27:40.138295 2709 reconciler.go:26] "Reconciler: start to sync state" Nov 24 00:27:40.138626 kubelet[2709]: E1124 00:27:40.138604 2709 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 24 00:27:40.139953 kubelet[2709]: I1124 00:27:40.139930 2709 factory.go:221] Registration of the systemd container factory successfully Nov 24 00:27:40.140030 kubelet[2709]: I1124 00:27:40.140007 2709 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 24 00:27:40.142010 kubelet[2709]: I1124 00:27:40.141876 2709 factory.go:221] Registration of the containerd container factory successfully Nov 24 00:27:40.142461 kubelet[2709]: I1124 00:27:40.142443 2709 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 00:27:40.143859 kubelet[2709]: I1124 00:27:40.143845 2709 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 00:27:40.144129 kubelet[2709]: I1124 00:27:40.144116 2709 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 24 00:27:40.144197 kubelet[2709]: I1124 00:27:40.144187 2709 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 24 00:27:40.144240 kubelet[2709]: I1124 00:27:40.144233 2709 kubelet.go:2382] "Starting kubelet main sync loop" Nov 24 00:27:40.144340 kubelet[2709]: E1124 00:27:40.144325 2709 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 00:27:40.196728 kubelet[2709]: I1124 00:27:40.196466 2709 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 24 00:27:40.196728 kubelet[2709]: I1124 00:27:40.196481 2709 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 24 00:27:40.196728 kubelet[2709]: I1124 00:27:40.196497 2709 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:27:40.196728 kubelet[2709]: I1124 00:27:40.196616 2709 state_mem.go:88] "Updated default CPUSet" cpuSet="" Nov 24 00:27:40.196728 kubelet[2709]: I1124 00:27:40.196626 2709 state_mem.go:96] "Updated CPUSet assignments" assignments={} Nov 24 00:27:40.196728 kubelet[2709]: I1124 00:27:40.196641 2709 policy_none.go:49] "None policy: Start" Nov 24 00:27:40.196728 kubelet[2709]: I1124 00:27:40.196649 2709 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 24 00:27:40.196728 kubelet[2709]: I1124 00:27:40.196659 2709 state_mem.go:35] "Initializing new in-memory state store" Nov 24 00:27:40.197266 kubelet[2709]: I1124 00:27:40.197253 2709 state_mem.go:75] "Updated machine memory state" Nov 24 00:27:40.201546 kubelet[2709]: I1124 00:27:40.201520 2709 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 00:27:40.201689 kubelet[2709]: I1124 00:27:40.201669 2709 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 00:27:40.201735 kubelet[2709]: I1124 00:27:40.201684 2709 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 00:27:40.202845 kubelet[2709]: I1124 00:27:40.202367 2709 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 00:27:40.207757 kubelet[2709]: E1124 00:27:40.207695 2709 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 24 00:27:40.245977 kubelet[2709]: I1124 00:27:40.245795 2709 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:40.246193 kubelet[2709]: I1124 00:27:40.245824 2709 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-172-237-154-236" Nov 24 00:27:40.246612 kubelet[2709]: I1124 00:27:40.246597 2709 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:40.311468 kubelet[2709]: I1124 00:27:40.310500 2709 kubelet_node_status.go:75] "Attempting to register node" node="172-237-154-236" Nov 24 00:27:40.320525 kubelet[2709]: I1124 00:27:40.320490 2709 kubelet_node_status.go:124] "Node was previously registered" node="172-237-154-236" Nov 24 00:27:40.320593 kubelet[2709]: I1124 00:27:40.320554 2709 kubelet_node_status.go:78] "Successfully registered node" node="172-237-154-236" Nov 24 00:27:40.339531 kubelet[2709]: I1124 00:27:40.339477 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/dcd4b3957f17b017ab6167b0ac63a70f-k8s-certs\") pod \"kube-apiserver-172-237-154-236\" (UID: \"dcd4b3957f17b017ab6167b0ac63a70f\") " pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:40.339531 kubelet[2709]: I1124 00:27:40.339525 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-flexvolume-dir\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:40.339655 kubelet[2709]: I1124 00:27:40.339544 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-k8s-certs\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:40.339655 kubelet[2709]: I1124 00:27:40.339562 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-kubeconfig\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:40.339655 kubelet[2709]: I1124 00:27:40.339595 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6d5965ce9dc5a2c239f1085a4e76621b-kubeconfig\") pod \"kube-scheduler-172-237-154-236\" (UID: \"6d5965ce9dc5a2c239f1085a4e76621b\") " pod="kube-system/kube-scheduler-172-237-154-236" Nov 24 00:27:40.339655 kubelet[2709]: I1124 00:27:40.339615 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/dcd4b3957f17b017ab6167b0ac63a70f-ca-certs\") pod \"kube-apiserver-172-237-154-236\" (UID: \"dcd4b3957f17b017ab6167b0ac63a70f\") " pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:40.339655 kubelet[2709]: I1124 00:27:40.339630 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/dcd4b3957f17b017ab6167b0ac63a70f-usr-share-ca-certificates\") pod \"kube-apiserver-172-237-154-236\" (UID: \"dcd4b3957f17b017ab6167b0ac63a70f\") " pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:40.339784 kubelet[2709]: I1124 00:27:40.339644 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-ca-certs\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:40.339784 kubelet[2709]: I1124 00:27:40.339674 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/973cb66da29b95ef423bd337cf05537d-usr-share-ca-certificates\") pod \"kube-controller-manager-172-237-154-236\" (UID: \"973cb66da29b95ef423bd337cf05537d\") " pod="kube-system/kube-controller-manager-172-237-154-236" Nov 24 00:27:40.555200 kubelet[2709]: E1124 00:27:40.555124 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:40.556017 kubelet[2709]: E1124 00:27:40.555472 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:40.556017 kubelet[2709]: E1124 00:27:40.555571 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:41.126361 kubelet[2709]: I1124 00:27:41.126309 2709 apiserver.go:52] "Watching apiserver" Nov 24 00:27:41.138409 kubelet[2709]: I1124 00:27:41.138366 2709 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 24 00:27:41.182021 kubelet[2709]: I1124 00:27:41.181989 2709 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:41.183226 kubelet[2709]: E1124 00:27:41.183199 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:41.183861 kubelet[2709]: E1124 00:27:41.183836 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:41.192193 kubelet[2709]: E1124 00:27:41.192154 2709 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-172-237-154-236\" already exists" pod="kube-system/kube-apiserver-172-237-154-236" Nov 24 00:27:41.192283 kubelet[2709]: E1124 00:27:41.192264 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:41.216052 kubelet[2709]: I1124 00:27:41.216007 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-172-237-154-236" podStartSLOduration=1.2159971139999999 podStartE2EDuration="1.215997114s" podCreationTimestamp="2025-11-24 00:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:41.20816665 +0000 UTC m=+1.154179948" watchObservedRunningTime="2025-11-24 00:27:41.215997114 +0000 UTC m=+1.162010422" Nov 24 00:27:41.222736 kubelet[2709]: I1124 00:27:41.222574 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-172-237-154-236" podStartSLOduration=1.222554967 podStartE2EDuration="1.222554967s" podCreationTimestamp="2025-11-24 00:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:41.221866217 +0000 UTC m=+1.167879525" watchObservedRunningTime="2025-11-24 00:27:41.222554967 +0000 UTC m=+1.168568265" Nov 24 00:27:41.222736 kubelet[2709]: I1124 00:27:41.222661 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-172-237-154-236" podStartSLOduration=1.222656677 podStartE2EDuration="1.222656677s" podCreationTimestamp="2025-11-24 00:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:41.216383384 +0000 UTC m=+1.162396682" watchObservedRunningTime="2025-11-24 00:27:41.222656677 +0000 UTC m=+1.168669975" Nov 24 00:27:42.183896 kubelet[2709]: E1124 00:27:42.183839 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:42.184791 kubelet[2709]: E1124 00:27:42.184762 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:43.055996 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Nov 24 00:27:45.222172 kubelet[2709]: I1124 00:27:45.222129 2709 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Nov 24 00:27:45.222907 containerd[1559]: time="2025-11-24T00:27:45.222848276Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Nov 24 00:27:45.223258 kubelet[2709]: I1124 00:27:45.223044 2709 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Nov 24 00:27:46.047918 kubelet[2709]: I1124 00:27:46.047858 2709 status_manager.go:890] "Failed to get status for pod" podUID="2294a018-4149-4a8b-92a7-3069e295bebd" pod="kube-system/kube-proxy-hbpfj" err="pods \"kube-proxy-hbpfj\" is forbidden: User \"system:node:172-237-154-236\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node '172-237-154-236' and this object" Nov 24 00:27:46.060785 systemd[1]: Created slice kubepods-besteffort-pod2294a018_4149_4a8b_92a7_3069e295bebd.slice - libcontainer container kubepods-besteffort-pod2294a018_4149_4a8b_92a7_3069e295bebd.slice. Nov 24 00:27:46.079144 kubelet[2709]: I1124 00:27:46.079105 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/2294a018-4149-4a8b-92a7-3069e295bebd-kube-proxy\") pod \"kube-proxy-hbpfj\" (UID: \"2294a018-4149-4a8b-92a7-3069e295bebd\") " pod="kube-system/kube-proxy-hbpfj" Nov 24 00:27:46.079415 kubelet[2709]: I1124 00:27:46.079308 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2294a018-4149-4a8b-92a7-3069e295bebd-xtables-lock\") pod \"kube-proxy-hbpfj\" (UID: \"2294a018-4149-4a8b-92a7-3069e295bebd\") " pod="kube-system/kube-proxy-hbpfj" Nov 24 00:27:46.079627 kubelet[2709]: I1124 00:27:46.079333 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2294a018-4149-4a8b-92a7-3069e295bebd-lib-modules\") pod \"kube-proxy-hbpfj\" (UID: \"2294a018-4149-4a8b-92a7-3069e295bebd\") " pod="kube-system/kube-proxy-hbpfj" Nov 24 00:27:46.079804 kubelet[2709]: I1124 00:27:46.079700 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpw5t\" (UniqueName: \"kubernetes.io/projected/2294a018-4149-4a8b-92a7-3069e295bebd-kube-api-access-gpw5t\") pod \"kube-proxy-hbpfj\" (UID: \"2294a018-4149-4a8b-92a7-3069e295bebd\") " pod="kube-system/kube-proxy-hbpfj" Nov 24 00:27:46.338759 systemd[1]: Created slice kubepods-besteffort-poda6ace751_7cf7_4030_9a9a_675eeb89ec95.slice - libcontainer container kubepods-besteffort-poda6ace751_7cf7_4030_9a9a_675eeb89ec95.slice. Nov 24 00:27:46.377953 kubelet[2709]: E1124 00:27:46.377901 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:46.379171 containerd[1559]: time="2025-11-24T00:27:46.379103224Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-hbpfj,Uid:2294a018-4149-4a8b-92a7-3069e295bebd,Namespace:kube-system,Attempt:0,}" Nov 24 00:27:46.382429 kubelet[2709]: I1124 00:27:46.382097 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/a6ace751-7cf7-4030-9a9a-675eeb89ec95-var-lib-calico\") pod \"tigera-operator-7dcd859c48-qk8hn\" (UID: \"a6ace751-7cf7-4030-9a9a-675eeb89ec95\") " pod="tigera-operator/tigera-operator-7dcd859c48-qk8hn" Nov 24 00:27:46.382601 kubelet[2709]: I1124 00:27:46.382552 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lwpz\" (UniqueName: \"kubernetes.io/projected/a6ace751-7cf7-4030-9a9a-675eeb89ec95-kube-api-access-7lwpz\") pod \"tigera-operator-7dcd859c48-qk8hn\" (UID: \"a6ace751-7cf7-4030-9a9a-675eeb89ec95\") " pod="tigera-operator/tigera-operator-7dcd859c48-qk8hn" Nov 24 00:27:46.401451 containerd[1559]: time="2025-11-24T00:27:46.400638545Z" level=info msg="connecting to shim e0703b0111d5c99b35872aba6f3a005c8186e4575383c30c718f41cbe24f2b53" address="unix:///run/containerd/s/8848b47cb39f98daddeef0e9eb54a7a2d3f84472100c21743bcdaeed9e8017a7" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:27:46.432015 systemd[1]: Started cri-containerd-e0703b0111d5c99b35872aba6f3a005c8186e4575383c30c718f41cbe24f2b53.scope - libcontainer container e0703b0111d5c99b35872aba6f3a005c8186e4575383c30c718f41cbe24f2b53. Nov 24 00:27:46.463068 containerd[1559]: time="2025-11-24T00:27:46.463030096Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-hbpfj,Uid:2294a018-4149-4a8b-92a7-3069e295bebd,Namespace:kube-system,Attempt:0,} returns sandbox id \"e0703b0111d5c99b35872aba6f3a005c8186e4575383c30c718f41cbe24f2b53\"" Nov 24 00:27:46.464449 kubelet[2709]: E1124 00:27:46.464420 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:46.468578 containerd[1559]: time="2025-11-24T00:27:46.468520159Z" level=info msg="CreateContainer within sandbox \"e0703b0111d5c99b35872aba6f3a005c8186e4575383c30c718f41cbe24f2b53\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Nov 24 00:27:46.484511 containerd[1559]: time="2025-11-24T00:27:46.484040016Z" level=info msg="Container f0fcb7c7c9a7c9fe7ef688bc95a598bcc98f77d1e9b9436d87cca0aab3b57370: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:27:46.494533 containerd[1559]: time="2025-11-24T00:27:46.494509462Z" level=info msg="CreateContainer within sandbox \"e0703b0111d5c99b35872aba6f3a005c8186e4575383c30c718f41cbe24f2b53\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"f0fcb7c7c9a7c9fe7ef688bc95a598bcc98f77d1e9b9436d87cca0aab3b57370\"" Nov 24 00:27:46.495550 containerd[1559]: time="2025-11-24T00:27:46.495384952Z" level=info msg="StartContainer for \"f0fcb7c7c9a7c9fe7ef688bc95a598bcc98f77d1e9b9436d87cca0aab3b57370\"" Nov 24 00:27:46.497726 containerd[1559]: time="2025-11-24T00:27:46.497675163Z" level=info msg="connecting to shim f0fcb7c7c9a7c9fe7ef688bc95a598bcc98f77d1e9b9436d87cca0aab3b57370" address="unix:///run/containerd/s/8848b47cb39f98daddeef0e9eb54a7a2d3f84472100c21743bcdaeed9e8017a7" protocol=ttrpc version=3 Nov 24 00:27:46.518843 systemd[1]: Started cri-containerd-f0fcb7c7c9a7c9fe7ef688bc95a598bcc98f77d1e9b9436d87cca0aab3b57370.scope - libcontainer container f0fcb7c7c9a7c9fe7ef688bc95a598bcc98f77d1e9b9436d87cca0aab3b57370. Nov 24 00:27:46.597468 containerd[1559]: time="2025-11-24T00:27:46.597175953Z" level=info msg="StartContainer for \"f0fcb7c7c9a7c9fe7ef688bc95a598bcc98f77d1e9b9436d87cca0aab3b57370\" returns successfully" Nov 24 00:27:46.643339 containerd[1559]: time="2025-11-24T00:27:46.643268046Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-qk8hn,Uid:a6ace751-7cf7-4030-9a9a-675eeb89ec95,Namespace:tigera-operator,Attempt:0,}" Nov 24 00:27:46.660277 containerd[1559]: time="2025-11-24T00:27:46.660093284Z" level=info msg="connecting to shim e6228dd6e931d062b3334bb6ca47b6b76b0b61571602320732713ba3561d854d" address="unix:///run/containerd/s/eba8914c1f9f2932a3e066781ae4fdc66b9dc44819d22d69317bc842f3f2102a" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:27:46.690863 systemd[1]: Started cri-containerd-e6228dd6e931d062b3334bb6ca47b6b76b0b61571602320732713ba3561d854d.scope - libcontainer container e6228dd6e931d062b3334bb6ca47b6b76b0b61571602320732713ba3561d854d. Nov 24 00:27:46.758885 containerd[1559]: time="2025-11-24T00:27:46.758419193Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-qk8hn,Uid:a6ace751-7cf7-4030-9a9a-675eeb89ec95,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"e6228dd6e931d062b3334bb6ca47b6b76b0b61571602320732713ba3561d854d\"" Nov 24 00:27:46.761923 containerd[1559]: time="2025-11-24T00:27:46.761903565Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Nov 24 00:27:47.200522 kubelet[2709]: E1124 00:27:47.200276 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:47.726242 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3854588815.mount: Deactivated successfully. Nov 24 00:27:48.010867 kubelet[2709]: E1124 00:27:48.010814 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:48.030559 kubelet[2709]: I1124 00:27:48.030232 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-hbpfj" podStartSLOduration=2.030211379 podStartE2EDuration="2.030211379s" podCreationTimestamp="2025-11-24 00:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:27:47.209021139 +0000 UTC m=+7.155034437" watchObservedRunningTime="2025-11-24 00:27:48.030211379 +0000 UTC m=+7.976224677" Nov 24 00:27:48.204202 kubelet[2709]: E1124 00:27:48.204145 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:48.405979 kubelet[2709]: E1124 00:27:48.405818 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:48.843725 kubelet[2709]: E1124 00:27:48.843675 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:49.021404 containerd[1559]: time="2025-11-24T00:27:49.021344124Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:49.022107 containerd[1559]: time="2025-11-24T00:27:49.022047954Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=25061691" Nov 24 00:27:49.022782 containerd[1559]: time="2025-11-24T00:27:49.022756875Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:49.024373 containerd[1559]: time="2025-11-24T00:27:49.024337096Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:27:49.025136 containerd[1559]: time="2025-11-24T00:27:49.024859386Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 2.262839001s" Nov 24 00:27:49.025136 containerd[1559]: time="2025-11-24T00:27:49.024884246Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Nov 24 00:27:49.027910 containerd[1559]: time="2025-11-24T00:27:49.027844197Z" level=info msg="CreateContainer within sandbox \"e6228dd6e931d062b3334bb6ca47b6b76b0b61571602320732713ba3561d854d\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Nov 24 00:27:49.038752 containerd[1559]: time="2025-11-24T00:27:49.035373831Z" level=info msg="Container cfa3c0255207c4690ef4db6c80ddb8e34db1204e65637560e139d8890080b1f7: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:27:49.038500 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount103356012.mount: Deactivated successfully. Nov 24 00:27:49.042351 containerd[1559]: time="2025-11-24T00:27:49.042310755Z" level=info msg="CreateContainer within sandbox \"e6228dd6e931d062b3334bb6ca47b6b76b0b61571602320732713ba3561d854d\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"cfa3c0255207c4690ef4db6c80ddb8e34db1204e65637560e139d8890080b1f7\"" Nov 24 00:27:49.043393 containerd[1559]: time="2025-11-24T00:27:49.042891715Z" level=info msg="StartContainer for \"cfa3c0255207c4690ef4db6c80ddb8e34db1204e65637560e139d8890080b1f7\"" Nov 24 00:27:49.043875 containerd[1559]: time="2025-11-24T00:27:49.043855105Z" level=info msg="connecting to shim cfa3c0255207c4690ef4db6c80ddb8e34db1204e65637560e139d8890080b1f7" address="unix:///run/containerd/s/eba8914c1f9f2932a3e066781ae4fdc66b9dc44819d22d69317bc842f3f2102a" protocol=ttrpc version=3 Nov 24 00:27:49.066842 systemd[1]: Started cri-containerd-cfa3c0255207c4690ef4db6c80ddb8e34db1204e65637560e139d8890080b1f7.scope - libcontainer container cfa3c0255207c4690ef4db6c80ddb8e34db1204e65637560e139d8890080b1f7. Nov 24 00:27:49.103119 containerd[1559]: time="2025-11-24T00:27:49.102639415Z" level=info msg="StartContainer for \"cfa3c0255207c4690ef4db6c80ddb8e34db1204e65637560e139d8890080b1f7\" returns successfully" Nov 24 00:27:49.206742 kubelet[2709]: E1124 00:27:49.206344 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:49.208862 kubelet[2709]: E1124 00:27:49.206605 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:50.211321 kubelet[2709]: E1124 00:27:50.210518 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:27:52.033261 systemd-timesyncd[1458]: Contacted time server [2600:3c01::f03c:93ff:fe5b:8a7d]:123 (2.flatcar.pool.ntp.org). Nov 24 00:27:52.033329 systemd-timesyncd[1458]: Initial clock synchronization to Mon 2025-11-24 00:27:52.033015 UTC. Nov 24 00:27:52.034955 systemd-resolved[1440]: Clock change detected. Flushing caches. Nov 24 00:27:55.330338 sudo[1795]: pam_unix(sudo:session): session closed for user root Nov 24 00:27:55.387946 sshd[1794]: Connection closed by 139.178.68.195 port 41572 Nov 24 00:27:55.387988 sshd-session[1791]: pam_unix(sshd:session): session closed for user core Nov 24 00:27:55.395380 systemd[1]: sshd@6-172.237.154.236:22-139.178.68.195:41572.service: Deactivated successfully. Nov 24 00:27:55.398828 systemd[1]: session-7.scope: Deactivated successfully. Nov 24 00:27:55.400106 systemd[1]: session-7.scope: Consumed 3.712s CPU time, 229.8M memory peak. Nov 24 00:27:55.403056 systemd-logind[1525]: Session 7 logged out. Waiting for processes to exit. Nov 24 00:27:55.406737 systemd-logind[1525]: Removed session 7. Nov 24 00:27:58.215961 update_engine[1530]: I20251124 00:27:58.215408 1530 update_attempter.cc:509] Updating boot flags... Nov 24 00:27:59.979030 kubelet[2709]: I1124 00:27:59.978940 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-qk8hn" podStartSLOduration=11.713642748 podStartE2EDuration="13.97887498s" podCreationTimestamp="2025-11-24 00:27:46 +0000 UTC" firstStartedPulling="2025-11-24 00:27:46.760445184 +0000 UTC m=+6.706458482" lastFinishedPulling="2025-11-24 00:27:49.025677416 +0000 UTC m=+8.971690714" observedRunningTime="2025-11-24 00:27:49.225029686 +0000 UTC m=+9.171042984" watchObservedRunningTime="2025-11-24 00:27:59.97887498 +0000 UTC m=+19.314683582" Nov 24 00:27:59.992040 systemd[1]: Created slice kubepods-besteffort-pod46de076c_fe66_4baa_9454_840ae067c054.slice - libcontainer container kubepods-besteffort-pod46de076c_fe66_4baa_9454_840ae067c054.slice. Nov 24 00:28:00.073934 kubelet[2709]: I1124 00:28:00.073862 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmlxg\" (UniqueName: \"kubernetes.io/projected/46de076c-fe66-4baa-9454-840ae067c054-kube-api-access-nmlxg\") pod \"calico-typha-6fd69ffcfd-vv7bc\" (UID: \"46de076c-fe66-4baa-9454-840ae067c054\") " pod="calico-system/calico-typha-6fd69ffcfd-vv7bc" Nov 24 00:28:00.074058 kubelet[2709]: I1124 00:28:00.073944 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/46de076c-fe66-4baa-9454-840ae067c054-typha-certs\") pod \"calico-typha-6fd69ffcfd-vv7bc\" (UID: \"46de076c-fe66-4baa-9454-840ae067c054\") " pod="calico-system/calico-typha-6fd69ffcfd-vv7bc" Nov 24 00:28:00.074058 kubelet[2709]: I1124 00:28:00.073968 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/46de076c-fe66-4baa-9454-840ae067c054-tigera-ca-bundle\") pod \"calico-typha-6fd69ffcfd-vv7bc\" (UID: \"46de076c-fe66-4baa-9454-840ae067c054\") " pod="calico-system/calico-typha-6fd69ffcfd-vv7bc" Nov 24 00:28:00.172390 systemd[1]: Created slice kubepods-besteffort-pod6c1ee368_2405_4063_8df7_ee001de64885.slice - libcontainer container kubepods-besteffort-pod6c1ee368_2405_4063_8df7_ee001de64885.slice. Nov 24 00:28:00.175034 kubelet[2709]: I1124 00:28:00.175009 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-xtables-lock\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.176103 kubelet[2709]: I1124 00:28:00.175154 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-lib-modules\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.176356 kubelet[2709]: I1124 00:28:00.176341 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-var-run-calico\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.176450 kubelet[2709]: I1124 00:28:00.176425 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-cni-log-dir\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.176554 kubelet[2709]: I1124 00:28:00.176537 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-flexvol-driver-host\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.177566 kubelet[2709]: I1124 00:28:00.177321 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-policysync\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.179133 kubelet[2709]: I1124 00:28:00.179089 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-var-lib-calico\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.179370 kubelet[2709]: I1124 00:28:00.179297 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj8rl\" (UniqueName: \"kubernetes.io/projected/6c1ee368-2405-4063-8df7-ee001de64885-kube-api-access-bj8rl\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.180022 kubelet[2709]: I1124 00:28:00.179449 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/6c1ee368-2405-4063-8df7-ee001de64885-node-certs\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.180022 kubelet[2709]: I1124 00:28:00.179483 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c1ee368-2405-4063-8df7-ee001de64885-tigera-ca-bundle\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.180022 kubelet[2709]: I1124 00:28:00.179504 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-cni-net-dir\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.180022 kubelet[2709]: I1124 00:28:00.179518 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/6c1ee368-2405-4063-8df7-ee001de64885-cni-bin-dir\") pod \"calico-node-pvs72\" (UID: \"6c1ee368-2405-4063-8df7-ee001de64885\") " pod="calico-system/calico-node-pvs72" Nov 24 00:28:00.282030 kubelet[2709]: E1124 00:28:00.281562 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.282165 kubelet[2709]: W1124 00:28:00.282146 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.283104 kubelet[2709]: E1124 00:28:00.283065 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.283500 kubelet[2709]: E1124 00:28:00.283349 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.283500 kubelet[2709]: W1124 00:28:00.283360 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.283779 kubelet[2709]: E1124 00:28:00.283653 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.283945 kubelet[2709]: E1124 00:28:00.283934 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.284025 kubelet[2709]: W1124 00:28:00.283995 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.284781 kubelet[2709]: E1124 00:28:00.284422 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.287987 kubelet[2709]: E1124 00:28:00.287957 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.287987 kubelet[2709]: W1124 00:28:00.287980 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.288063 kubelet[2709]: E1124 00:28:00.288000 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.293987 kubelet[2709]: E1124 00:28:00.293945 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.293987 kubelet[2709]: W1124 00:28:00.293967 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.293987 kubelet[2709]: E1124 00:28:00.293988 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.296933 kubelet[2709]: E1124 00:28:00.294198 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.296933 kubelet[2709]: W1124 00:28:00.294210 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.296933 kubelet[2709]: E1124 00:28:00.294218 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.297348 kubelet[2709]: E1124 00:28:00.297322 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:00.297699 containerd[1559]: time="2025-11-24T00:28:00.297664389Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6fd69ffcfd-vv7bc,Uid:46de076c-fe66-4baa-9454-840ae067c054,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:00.315734 containerd[1559]: time="2025-11-24T00:28:00.315550778Z" level=info msg="connecting to shim 5a70cf5d23162deee264b5d4da76b6608f478f88190a639055201de6960ebca9" address="unix:///run/containerd/s/9e16ee647e5e6f5c66dbfbe711e15b0126d7e7e7b9857be4b4016ae1dd56db63" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:00.339183 systemd[1]: Started cri-containerd-5a70cf5d23162deee264b5d4da76b6608f478f88190a639055201de6960ebca9.scope - libcontainer container 5a70cf5d23162deee264b5d4da76b6608f478f88190a639055201de6960ebca9. Nov 24 00:28:00.376231 kubelet[2709]: E1124 00:28:00.376187 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:00.417167 containerd[1559]: time="2025-11-24T00:28:00.417085149Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6fd69ffcfd-vv7bc,Uid:46de076c-fe66-4baa-9454-840ae067c054,Namespace:calico-system,Attempt:0,} returns sandbox id \"5a70cf5d23162deee264b5d4da76b6608f478f88190a639055201de6960ebca9\"" Nov 24 00:28:00.417688 kubelet[2709]: E1124 00:28:00.417665 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:00.419908 containerd[1559]: time="2025-11-24T00:28:00.419771450Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Nov 24 00:28:00.471547 kubelet[2709]: E1124 00:28:00.471521 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.471547 kubelet[2709]: W1124 00:28:00.471545 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.471677 kubelet[2709]: E1124 00:28:00.471561 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.471769 kubelet[2709]: E1124 00:28:00.471752 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.471769 kubelet[2709]: W1124 00:28:00.471766 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.471987 kubelet[2709]: E1124 00:28:00.471774 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.472045 kubelet[2709]: E1124 00:28:00.472022 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.472045 kubelet[2709]: W1124 00:28:00.472038 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.472115 kubelet[2709]: E1124 00:28:00.472046 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.472329 kubelet[2709]: E1124 00:28:00.472289 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.472329 kubelet[2709]: W1124 00:28:00.472299 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.472329 kubelet[2709]: E1124 00:28:00.472307 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.472552 kubelet[2709]: E1124 00:28:00.472534 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.472552 kubelet[2709]: W1124 00:28:00.472547 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.472632 kubelet[2709]: E1124 00:28:00.472556 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.472861 kubelet[2709]: E1124 00:28:00.472846 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.472861 kubelet[2709]: W1124 00:28:00.472857 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.472947 kubelet[2709]: E1124 00:28:00.472865 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.473123 kubelet[2709]: E1124 00:28:00.473107 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.473123 kubelet[2709]: W1124 00:28:00.473121 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.473207 kubelet[2709]: E1124 00:28:00.473132 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.473339 kubelet[2709]: E1124 00:28:00.473319 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.473339 kubelet[2709]: W1124 00:28:00.473330 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.473339 kubelet[2709]: E1124 00:28:00.473338 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.473713 kubelet[2709]: E1124 00:28:00.473556 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.473713 kubelet[2709]: W1124 00:28:00.473588 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.473713 kubelet[2709]: E1124 00:28:00.473607 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.473888 kubelet[2709]: E1124 00:28:00.473845 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.473888 kubelet[2709]: W1124 00:28:00.473856 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.473888 kubelet[2709]: E1124 00:28:00.473865 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.474393 kubelet[2709]: E1124 00:28:00.474074 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.474393 kubelet[2709]: W1124 00:28:00.474089 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.474393 kubelet[2709]: E1124 00:28:00.474103 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.474393 kubelet[2709]: E1124 00:28:00.474368 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.474393 kubelet[2709]: W1124 00:28:00.474376 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.474393 kubelet[2709]: E1124 00:28:00.474383 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.475094 kubelet[2709]: E1124 00:28:00.474598 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.475094 kubelet[2709]: W1124 00:28:00.474606 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.475094 kubelet[2709]: E1124 00:28:00.474613 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.475094 kubelet[2709]: E1124 00:28:00.474787 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.475094 kubelet[2709]: W1124 00:28:00.474793 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.475094 kubelet[2709]: E1124 00:28:00.474801 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.475094 kubelet[2709]: E1124 00:28:00.474993 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.475094 kubelet[2709]: W1124 00:28:00.475001 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.475094 kubelet[2709]: E1124 00:28:00.475008 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.476111 kubelet[2709]: E1124 00:28:00.475182 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.476111 kubelet[2709]: W1124 00:28:00.475190 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.476111 kubelet[2709]: E1124 00:28:00.475198 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.476111 kubelet[2709]: E1124 00:28:00.475375 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.476111 kubelet[2709]: W1124 00:28:00.475382 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.476111 kubelet[2709]: E1124 00:28:00.475389 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.476111 kubelet[2709]: E1124 00:28:00.475560 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.476111 kubelet[2709]: W1124 00:28:00.475567 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.476111 kubelet[2709]: E1124 00:28:00.475575 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.476111 kubelet[2709]: E1124 00:28:00.475817 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.476615 kubelet[2709]: W1124 00:28:00.475825 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.476615 kubelet[2709]: E1124 00:28:00.475854 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.476615 kubelet[2709]: E1124 00:28:00.476113 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.476615 kubelet[2709]: W1124 00:28:00.476121 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.476615 kubelet[2709]: E1124 00:28:00.476129 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.479363 kubelet[2709]: E1124 00:28:00.479341 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:00.479761 containerd[1559]: time="2025-11-24T00:28:00.479685560Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-pvs72,Uid:6c1ee368-2405-4063-8df7-ee001de64885,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:00.482477 kubelet[2709]: E1124 00:28:00.482432 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.482477 kubelet[2709]: W1124 00:28:00.482474 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.482546 kubelet[2709]: E1124 00:28:00.482484 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.482546 kubelet[2709]: I1124 00:28:00.482507 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/d80649f2-a0ab-46f8-9106-ed888d934bff-varrun\") pod \"csi-node-driver-5vjmh\" (UID: \"d80649f2-a0ab-46f8-9106-ed888d934bff\") " pod="calico-system/csi-node-driver-5vjmh" Nov 24 00:28:00.482797 kubelet[2709]: E1124 00:28:00.482770 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.482797 kubelet[2709]: W1124 00:28:00.482788 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.483401 kubelet[2709]: E1124 00:28:00.482819 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.483401 kubelet[2709]: I1124 00:28:00.482961 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d80649f2-a0ab-46f8-9106-ed888d934bff-kubelet-dir\") pod \"csi-node-driver-5vjmh\" (UID: \"d80649f2-a0ab-46f8-9106-ed888d934bff\") " pod="calico-system/csi-node-driver-5vjmh" Nov 24 00:28:00.483401 kubelet[2709]: E1124 00:28:00.483193 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.483401 kubelet[2709]: W1124 00:28:00.483388 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.483486 kubelet[2709]: E1124 00:28:00.483412 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.484206 kubelet[2709]: E1124 00:28:00.483628 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.484206 kubelet[2709]: W1124 00:28:00.483638 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.484206 kubelet[2709]: E1124 00:28:00.483679 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.484206 kubelet[2709]: E1124 00:28:00.483970 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.484206 kubelet[2709]: W1124 00:28:00.483978 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.484206 kubelet[2709]: E1124 00:28:00.484206 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.484865 kubelet[2709]: I1124 00:28:00.484225 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d80649f2-a0ab-46f8-9106-ed888d934bff-registration-dir\") pod \"csi-node-driver-5vjmh\" (UID: \"d80649f2-a0ab-46f8-9106-ed888d934bff\") " pod="calico-system/csi-node-driver-5vjmh" Nov 24 00:28:00.484865 kubelet[2709]: E1124 00:28:00.484588 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.484865 kubelet[2709]: W1124 00:28:00.484603 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.484865 kubelet[2709]: E1124 00:28:00.484693 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.485263 kubelet[2709]: E1124 00:28:00.485082 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.485263 kubelet[2709]: W1124 00:28:00.485091 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.485263 kubelet[2709]: E1124 00:28:00.485138 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.485901 kubelet[2709]: E1124 00:28:00.485400 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.485901 kubelet[2709]: W1124 00:28:00.485408 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.485901 kubelet[2709]: E1124 00:28:00.485425 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.485901 kubelet[2709]: I1124 00:28:00.485603 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brnt9\" (UniqueName: \"kubernetes.io/projected/d80649f2-a0ab-46f8-9106-ed888d934bff-kube-api-access-brnt9\") pod \"csi-node-driver-5vjmh\" (UID: \"d80649f2-a0ab-46f8-9106-ed888d934bff\") " pod="calico-system/csi-node-driver-5vjmh" Nov 24 00:28:00.485901 kubelet[2709]: E1124 00:28:00.485690 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.485901 kubelet[2709]: W1124 00:28:00.485699 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.485901 kubelet[2709]: E1124 00:28:00.485757 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.486094 kubelet[2709]: E1124 00:28:00.486046 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.486094 kubelet[2709]: W1124 00:28:00.486081 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.486146 kubelet[2709]: E1124 00:28:00.486098 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.487013 kubelet[2709]: E1124 00:28:00.486531 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.487013 kubelet[2709]: W1124 00:28:00.486567 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.487013 kubelet[2709]: E1124 00:28:00.486584 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.487013 kubelet[2709]: I1124 00:28:00.486598 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d80649f2-a0ab-46f8-9106-ed888d934bff-socket-dir\") pod \"csi-node-driver-5vjmh\" (UID: \"d80649f2-a0ab-46f8-9106-ed888d934bff\") " pod="calico-system/csi-node-driver-5vjmh" Nov 24 00:28:00.487013 kubelet[2709]: E1124 00:28:00.486906 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.487013 kubelet[2709]: W1124 00:28:00.486958 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.487436 kubelet[2709]: E1124 00:28:00.487092 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.487436 kubelet[2709]: E1124 00:28:00.487375 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.487436 kubelet[2709]: W1124 00:28:00.487383 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.487436 kubelet[2709]: E1124 00:28:00.487395 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.487953 kubelet[2709]: E1124 00:28:00.487862 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.487953 kubelet[2709]: W1124 00:28:00.487875 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.487953 kubelet[2709]: E1124 00:28:00.487883 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.488599 kubelet[2709]: E1124 00:28:00.488192 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.488599 kubelet[2709]: W1124 00:28:00.488200 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.488599 kubelet[2709]: E1124 00:28:00.488208 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.508510 containerd[1559]: time="2025-11-24T00:28:00.508141115Z" level=info msg="connecting to shim 89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673" address="unix:///run/containerd/s/dc465aa4d2a3f40975a175965f23c6e15aef3c4794f2fa167bcc6014e7cd1389" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:00.550091 systemd[1]: Started cri-containerd-89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673.scope - libcontainer container 89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673. Nov 24 00:28:00.588176 kubelet[2709]: E1124 00:28:00.588137 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.588176 kubelet[2709]: W1124 00:28:00.588165 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.588442 kubelet[2709]: E1124 00:28:00.588210 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.588736 kubelet[2709]: E1124 00:28:00.588594 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.588736 kubelet[2709]: W1124 00:28:00.588611 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.588736 kubelet[2709]: E1124 00:28:00.588639 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.589044 kubelet[2709]: E1124 00:28:00.588861 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.589044 kubelet[2709]: W1124 00:28:00.588871 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.589044 kubelet[2709]: E1124 00:28:00.588884 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.589338 kubelet[2709]: E1124 00:28:00.589193 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.589338 kubelet[2709]: W1124 00:28:00.589204 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.589338 kubelet[2709]: E1124 00:28:00.589231 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.589691 kubelet[2709]: E1124 00:28:00.589564 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.589691 kubelet[2709]: W1124 00:28:00.589575 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.589691 kubelet[2709]: E1124 00:28:00.589587 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.590001 kubelet[2709]: E1124 00:28:00.589980 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.590198 kubelet[2709]: W1124 00:28:00.590138 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.590198 kubelet[2709]: E1124 00:28:00.590167 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.590951 kubelet[2709]: E1124 00:28:00.590824 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.591046 kubelet[2709]: W1124 00:28:00.591020 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.591238 kubelet[2709]: E1124 00:28:00.591208 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.591824 kubelet[2709]: E1124 00:28:00.591777 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.591824 kubelet[2709]: W1124 00:28:00.591789 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.591824 kubelet[2709]: E1124 00:28:00.591804 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.592311 kubelet[2709]: E1124 00:28:00.592131 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.592311 kubelet[2709]: W1124 00:28:00.592141 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.592311 kubelet[2709]: E1124 00:28:00.592153 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.592458 kubelet[2709]: E1124 00:28:00.592446 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.592629 kubelet[2709]: W1124 00:28:00.592499 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.592629 kubelet[2709]: E1124 00:28:00.592585 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.593044 kubelet[2709]: E1124 00:28:00.592972 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.593044 kubelet[2709]: W1124 00:28:00.592984 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.593203 kubelet[2709]: E1124 00:28:00.593127 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.593968 kubelet[2709]: E1124 00:28:00.593330 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.594142 kubelet[2709]: W1124 00:28:00.594020 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.594142 kubelet[2709]: E1124 00:28:00.594078 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.594269 kubelet[2709]: E1124 00:28:00.594258 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.594322 kubelet[2709]: W1124 00:28:00.594312 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.594437 kubelet[2709]: E1124 00:28:00.594427 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.594677 kubelet[2709]: E1124 00:28:00.594666 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.594770 kubelet[2709]: W1124 00:28:00.594758 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.594906 kubelet[2709]: E1124 00:28:00.594896 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.595367 kubelet[2709]: E1124 00:28:00.595355 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.595428 kubelet[2709]: W1124 00:28:00.595418 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.595513 kubelet[2709]: E1124 00:28:00.595501 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.595732 kubelet[2709]: E1124 00:28:00.595721 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.595788 kubelet[2709]: W1124 00:28:00.595778 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.595955 kubelet[2709]: E1124 00:28:00.595943 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.596354 kubelet[2709]: E1124 00:28:00.596342 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.596414 kubelet[2709]: W1124 00:28:00.596404 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.596594 kubelet[2709]: E1124 00:28:00.596581 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.596988 kubelet[2709]: E1124 00:28:00.596977 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.597053 kubelet[2709]: W1124 00:28:00.597041 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.597189 kubelet[2709]: E1124 00:28:00.597178 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.598136 kubelet[2709]: E1124 00:28:00.597986 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.598136 kubelet[2709]: W1124 00:28:00.597998 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.598136 kubelet[2709]: E1124 00:28:00.598099 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.598460 kubelet[2709]: E1124 00:28:00.598294 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.598460 kubelet[2709]: W1124 00:28:00.598305 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.598460 kubelet[2709]: E1124 00:28:00.598401 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.598713 kubelet[2709]: E1124 00:28:00.598702 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.598768 kubelet[2709]: W1124 00:28:00.598758 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.598887 kubelet[2709]: E1124 00:28:00.598876 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.599443 kubelet[2709]: E1124 00:28:00.599432 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.599615 kubelet[2709]: W1124 00:28:00.599527 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.599743 kubelet[2709]: E1124 00:28:00.599673 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.600315 kubelet[2709]: E1124 00:28:00.600250 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.600315 kubelet[2709]: W1124 00:28:00.600260 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.600608 kubelet[2709]: E1124 00:28:00.600555 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.601262 kubelet[2709]: E1124 00:28:00.601170 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.601262 kubelet[2709]: W1124 00:28:00.601182 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.601491 kubelet[2709]: E1124 00:28:00.601415 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.601707 kubelet[2709]: E1124 00:28:00.601695 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.601970 kubelet[2709]: W1124 00:28:00.601772 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.601970 kubelet[2709]: E1124 00:28:00.601784 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:00.603026 containerd[1559]: time="2025-11-24T00:28:00.602947162Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-pvs72,Uid:6c1ee368-2405-4063-8df7-ee001de64885,Namespace:calico-system,Attempt:0,} returns sandbox id \"89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673\"" Nov 24 00:28:00.603812 kubelet[2709]: E1124 00:28:00.603762 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:00.612185 kubelet[2709]: E1124 00:28:00.612143 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:00.612185 kubelet[2709]: W1124 00:28:00.612155 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:00.612185 kubelet[2709]: E1124 00:28:00.612165 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.561209 containerd[1559]: time="2025-11-24T00:28:01.561137631Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:01.562354 containerd[1559]: time="2025-11-24T00:28:01.562299461Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=35234628" Nov 24 00:28:01.563192 containerd[1559]: time="2025-11-24T00:28:01.563146042Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:01.568441 containerd[1559]: time="2025-11-24T00:28:01.568371964Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:01.570613 containerd[1559]: time="2025-11-24T00:28:01.570506445Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 1.150688605s" Nov 24 00:28:01.570613 containerd[1559]: time="2025-11-24T00:28:01.570532115Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Nov 24 00:28:01.573185 containerd[1559]: time="2025-11-24T00:28:01.573152407Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Nov 24 00:28:01.589957 containerd[1559]: time="2025-11-24T00:28:01.589882935Z" level=info msg="CreateContainer within sandbox \"5a70cf5d23162deee264b5d4da76b6608f478f88190a639055201de6960ebca9\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Nov 24 00:28:01.599937 containerd[1559]: time="2025-11-24T00:28:01.597183119Z" level=info msg="Container 27dfe439483a03db08388bc87160f3609e6ba0db098ba110d41e11742697fb6d: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:28:01.599928 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4291610901.mount: Deactivated successfully. Nov 24 00:28:01.607384 containerd[1559]: time="2025-11-24T00:28:01.607345284Z" level=info msg="CreateContainer within sandbox \"5a70cf5d23162deee264b5d4da76b6608f478f88190a639055201de6960ebca9\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"27dfe439483a03db08388bc87160f3609e6ba0db098ba110d41e11742697fb6d\"" Nov 24 00:28:01.608707 containerd[1559]: time="2025-11-24T00:28:01.607807304Z" level=info msg="StartContainer for \"27dfe439483a03db08388bc87160f3609e6ba0db098ba110d41e11742697fb6d\"" Nov 24 00:28:01.609611 containerd[1559]: time="2025-11-24T00:28:01.609587505Z" level=info msg="connecting to shim 27dfe439483a03db08388bc87160f3609e6ba0db098ba110d41e11742697fb6d" address="unix:///run/containerd/s/9e16ee647e5e6f5c66dbfbe711e15b0126d7e7e7b9857be4b4016ae1dd56db63" protocol=ttrpc version=3 Nov 24 00:28:01.631173 systemd[1]: Started cri-containerd-27dfe439483a03db08388bc87160f3609e6ba0db098ba110d41e11742697fb6d.scope - libcontainer container 27dfe439483a03db08388bc87160f3609e6ba0db098ba110d41e11742697fb6d. Nov 24 00:28:01.703883 containerd[1559]: time="2025-11-24T00:28:01.703803452Z" level=info msg="StartContainer for \"27dfe439483a03db08388bc87160f3609e6ba0db098ba110d41e11742697fb6d\" returns successfully" Nov 24 00:28:01.755519 kubelet[2709]: E1124 00:28:01.755462 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:01.845124 kubelet[2709]: E1124 00:28:01.843685 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:01.883749 kubelet[2709]: E1124 00:28:01.883724 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.883886 kubelet[2709]: W1124 00:28:01.883873 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.883992 kubelet[2709]: E1124 00:28:01.883963 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.884402 kubelet[2709]: E1124 00:28:01.884391 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.884491 kubelet[2709]: W1124 00:28:01.884442 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.884491 kubelet[2709]: E1124 00:28:01.884454 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.884771 kubelet[2709]: E1124 00:28:01.884712 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.884771 kubelet[2709]: W1124 00:28:01.884724 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.884771 kubelet[2709]: E1124 00:28:01.884732 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.886644 kubelet[2709]: E1124 00:28:01.886205 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.886644 kubelet[2709]: W1124 00:28:01.886216 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.886644 kubelet[2709]: E1124 00:28:01.886226 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.886813 kubelet[2709]: E1124 00:28:01.886802 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.886860 kubelet[2709]: W1124 00:28:01.886850 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.886904 kubelet[2709]: E1124 00:28:01.886894 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.887172 kubelet[2709]: E1124 00:28:01.887162 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.887275 kubelet[2709]: W1124 00:28:01.887225 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.887275 kubelet[2709]: E1124 00:28:01.887238 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.887594 kubelet[2709]: E1124 00:28:01.887582 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.887739 kubelet[2709]: W1124 00:28:01.887642 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.887739 kubelet[2709]: E1124 00:28:01.887655 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.887981 kubelet[2709]: E1124 00:28:01.887969 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.888145 kubelet[2709]: W1124 00:28:01.888132 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.888237 kubelet[2709]: E1124 00:28:01.888212 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.889223 kubelet[2709]: E1124 00:28:01.889163 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.889223 kubelet[2709]: W1124 00:28:01.889175 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.889223 kubelet[2709]: E1124 00:28:01.889184 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.889510 kubelet[2709]: E1124 00:28:01.889456 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.889510 kubelet[2709]: W1124 00:28:01.889466 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.889510 kubelet[2709]: E1124 00:28:01.889473 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.889832 kubelet[2709]: E1124 00:28:01.889767 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.889832 kubelet[2709]: W1124 00:28:01.889778 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.889832 kubelet[2709]: E1124 00:28:01.889785 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.890990 kubelet[2709]: E1124 00:28:01.890893 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.890990 kubelet[2709]: W1124 00:28:01.890904 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.890990 kubelet[2709]: E1124 00:28:01.890942 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.891260 kubelet[2709]: E1124 00:28:01.891250 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.891309 kubelet[2709]: W1124 00:28:01.891300 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.891392 kubelet[2709]: E1124 00:28:01.891350 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.891694 kubelet[2709]: E1124 00:28:01.891683 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.891954 kubelet[2709]: W1124 00:28:01.891941 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.892074 kubelet[2709]: E1124 00:28:01.891993 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.892665 kubelet[2709]: E1124 00:28:01.892641 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.892724 kubelet[2709]: W1124 00:28:01.892714 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.892802 kubelet[2709]: E1124 00:28:01.892757 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.901626 kubelet[2709]: E1124 00:28:01.901582 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.901626 kubelet[2709]: W1124 00:28:01.901596 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.901626 kubelet[2709]: E1124 00:28:01.901607 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.902054 kubelet[2709]: E1124 00:28:01.902030 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.902054 kubelet[2709]: W1124 00:28:01.902041 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.902190 kubelet[2709]: E1124 00:28:01.902135 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.902447 kubelet[2709]: E1124 00:28:01.902425 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.902447 kubelet[2709]: W1124 00:28:01.902434 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.902570 kubelet[2709]: E1124 00:28:01.902514 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.902811 kubelet[2709]: E1124 00:28:01.902789 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.902811 kubelet[2709]: W1124 00:28:01.902799 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.902933 kubelet[2709]: E1124 00:28:01.902871 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.903143 kubelet[2709]: E1124 00:28:01.903121 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.903143 kubelet[2709]: W1124 00:28:01.903131 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.903291 kubelet[2709]: E1124 00:28:01.903271 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.904094 kubelet[2709]: E1124 00:28:01.904070 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.904094 kubelet[2709]: W1124 00:28:01.904081 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.904423 kubelet[2709]: E1124 00:28:01.904401 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.904607 kubelet[2709]: E1124 00:28:01.904586 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.904607 kubelet[2709]: W1124 00:28:01.904595 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.906043 kubelet[2709]: E1124 00:28:01.906017 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.906251 kubelet[2709]: E1124 00:28:01.906228 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.906251 kubelet[2709]: W1124 00:28:01.906238 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.906675 kubelet[2709]: E1124 00:28:01.906652 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.906854 kubelet[2709]: E1124 00:28:01.906777 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.906854 kubelet[2709]: W1124 00:28:01.906787 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.907034 kubelet[2709]: E1124 00:28:01.907013 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.907133 kubelet[2709]: E1124 00:28:01.907123 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.907190 kubelet[2709]: W1124 00:28:01.907167 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.907641 kubelet[2709]: E1124 00:28:01.907420 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.907641 kubelet[2709]: W1124 00:28:01.907429 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.907641 kubelet[2709]: E1124 00:28:01.907437 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.907888 kubelet[2709]: E1124 00:28:01.907228 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.908258 kubelet[2709]: E1124 00:28:01.908216 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.909011 kubelet[2709]: W1124 00:28:01.908822 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.909011 kubelet[2709]: E1124 00:28:01.908848 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.909257 kubelet[2709]: E1124 00:28:01.909186 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.909257 kubelet[2709]: W1124 00:28:01.909244 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.909935 kubelet[2709]: E1124 00:28:01.909378 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.910401 kubelet[2709]: E1124 00:28:01.910358 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.910401 kubelet[2709]: W1124 00:28:01.910391 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.911581 kubelet[2709]: E1124 00:28:01.911547 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.912008 kubelet[2709]: E1124 00:28:01.911981 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.912008 kubelet[2709]: W1124 00:28:01.911999 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.912295 kubelet[2709]: E1124 00:28:01.912053 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.912461 kubelet[2709]: E1124 00:28:01.912437 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.912461 kubelet[2709]: W1124 00:28:01.912453 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.912506 kubelet[2709]: E1124 00:28:01.912463 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.914316 kubelet[2709]: E1124 00:28:01.914291 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.914316 kubelet[2709]: W1124 00:28:01.914310 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.914383 kubelet[2709]: E1124 00:28:01.914322 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:01.914564 kubelet[2709]: E1124 00:28:01.914542 2709 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:28:01.914564 kubelet[2709]: W1124 00:28:01.914558 2709 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:28:01.914607 kubelet[2709]: E1124 00:28:01.914567 2709 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:28:02.245751 containerd[1559]: time="2025-11-24T00:28:02.245695423Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:02.246388 containerd[1559]: time="2025-11-24T00:28:02.246362563Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=4446754" Nov 24 00:28:02.246953 containerd[1559]: time="2025-11-24T00:28:02.246884383Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:02.248183 containerd[1559]: time="2025-11-24T00:28:02.248161824Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:02.248941 containerd[1559]: time="2025-11-24T00:28:02.248703074Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 675.516327ms" Nov 24 00:28:02.248941 containerd[1559]: time="2025-11-24T00:28:02.248733664Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Nov 24 00:28:02.250801 containerd[1559]: time="2025-11-24T00:28:02.250780665Z" level=info msg="CreateContainer within sandbox \"89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Nov 24 00:28:02.259229 containerd[1559]: time="2025-11-24T00:28:02.259207940Z" level=info msg="Container e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:28:02.263554 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2570806245.mount: Deactivated successfully. Nov 24 00:28:02.271825 containerd[1559]: time="2025-11-24T00:28:02.271770526Z" level=info msg="CreateContainer within sandbox \"89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf\"" Nov 24 00:28:02.272705 containerd[1559]: time="2025-11-24T00:28:02.272679806Z" level=info msg="StartContainer for \"e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf\"" Nov 24 00:28:02.274053 containerd[1559]: time="2025-11-24T00:28:02.274019437Z" level=info msg="connecting to shim e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf" address="unix:///run/containerd/s/dc465aa4d2a3f40975a175965f23c6e15aef3c4794f2fa167bcc6014e7cd1389" protocol=ttrpc version=3 Nov 24 00:28:02.298267 systemd[1]: Started cri-containerd-e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf.scope - libcontainer container e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf. Nov 24 00:28:02.379567 containerd[1559]: time="2025-11-24T00:28:02.379520310Z" level=info msg="StartContainer for \"e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf\" returns successfully" Nov 24 00:28:02.398842 systemd[1]: cri-containerd-e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf.scope: Deactivated successfully. Nov 24 00:28:02.402547 containerd[1559]: time="2025-11-24T00:28:02.402508191Z" level=info msg="received container exit event container_id:\"e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf\" id:\"e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf\" pid:3400 exited_at:{seconds:1763944082 nanos:402106001}" Nov 24 00:28:02.429873 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e0a6189182195bea16a09878d2c30df8907999b62be45f45c895f911bf2a3faf-rootfs.mount: Deactivated successfully. Nov 24 00:28:02.848288 kubelet[2709]: I1124 00:28:02.848231 2709 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 00:28:02.849865 kubelet[2709]: E1124 00:28:02.848559 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:02.851529 kubelet[2709]: E1124 00:28:02.851454 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:02.851998 containerd[1559]: time="2025-11-24T00:28:02.851783616Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Nov 24 00:28:02.866593 kubelet[2709]: I1124 00:28:02.866538 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6fd69ffcfd-vv7bc" podStartSLOduration=2.713281347 podStartE2EDuration="3.866522453s" podCreationTimestamp="2025-11-24 00:27:59 +0000 UTC" firstStartedPulling="2025-11-24 00:28:00.41838641 +0000 UTC m=+19.754195012" lastFinishedPulling="2025-11-24 00:28:01.571627516 +0000 UTC m=+20.907436118" observedRunningTime="2025-11-24 00:28:01.876327328 +0000 UTC m=+21.212135930" watchObservedRunningTime="2025-11-24 00:28:02.866522453 +0000 UTC m=+22.202331055" Nov 24 00:28:03.755527 kubelet[2709]: E1124 00:28:03.755274 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:04.631666 containerd[1559]: time="2025-11-24T00:28:04.631608465Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:04.633004 containerd[1559]: time="2025-11-24T00:28:04.632482905Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70446859" Nov 24 00:28:04.633728 containerd[1559]: time="2025-11-24T00:28:04.633687926Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:04.636428 containerd[1559]: time="2025-11-24T00:28:04.636181987Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:04.636796 containerd[1559]: time="2025-11-24T00:28:04.636768317Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 1.784957021s" Nov 24 00:28:04.636968 containerd[1559]: time="2025-11-24T00:28:04.636795457Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Nov 24 00:28:04.640122 containerd[1559]: time="2025-11-24T00:28:04.640096129Z" level=info msg="CreateContainer within sandbox \"89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Nov 24 00:28:04.649768 containerd[1559]: time="2025-11-24T00:28:04.647112373Z" level=info msg="Container 8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:28:04.655282 containerd[1559]: time="2025-11-24T00:28:04.655238987Z" level=info msg="CreateContainer within sandbox \"89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c\"" Nov 24 00:28:04.655949 containerd[1559]: time="2025-11-24T00:28:04.655745227Z" level=info msg="StartContainer for \"8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c\"" Nov 24 00:28:04.657271 containerd[1559]: time="2025-11-24T00:28:04.657251128Z" level=info msg="connecting to shim 8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c" address="unix:///run/containerd/s/dc465aa4d2a3f40975a175965f23c6e15aef3c4794f2fa167bcc6014e7cd1389" protocol=ttrpc version=3 Nov 24 00:28:04.684081 systemd[1]: Started cri-containerd-8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c.scope - libcontainer container 8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c. Nov 24 00:28:04.791709 containerd[1559]: time="2025-11-24T00:28:04.791649445Z" level=info msg="StartContainer for \"8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c\" returns successfully" Nov 24 00:28:04.855811 kubelet[2709]: E1124 00:28:04.855775 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:05.297055 containerd[1559]: time="2025-11-24T00:28:05.297013617Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 24 00:28:05.299827 systemd[1]: cri-containerd-8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c.scope: Deactivated successfully. Nov 24 00:28:05.300612 systemd[1]: cri-containerd-8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c.scope: Consumed 530ms CPU time, 192.3M memory peak, 171.3M written to disk. Nov 24 00:28:05.303248 containerd[1559]: time="2025-11-24T00:28:05.303203150Z" level=info msg="received container exit event container_id:\"8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c\" id:\"8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c\" pid:3459 exited_at:{seconds:1763944085 nanos:302932660}" Nov 24 00:28:05.327104 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8dc644089ec423ef42135b45f29d38c4558ad1e01dea31d238ad70e032ada48c-rootfs.mount: Deactivated successfully. Nov 24 00:28:05.362025 kubelet[2709]: I1124 00:28:05.361999 2709 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Nov 24 00:28:05.399819 kubelet[2709]: W1124 00:28:05.399778 2709 reflector.go:569] object-"calico-system"/"whisker-backend-key-pair": failed to list *v1.Secret: secrets "whisker-backend-key-pair" is forbidden: User "system:node:172-237-154-236" cannot list resource "secrets" in API group "" in the namespace "calico-system": no relationship found between node '172-237-154-236' and this object Nov 24 00:28:05.401882 kubelet[2709]: E1124 00:28:05.401848 2709 reflector.go:166] "Unhandled Error" err="object-\"calico-system\"/\"whisker-backend-key-pair\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"whisker-backend-key-pair\" is forbidden: User \"system:node:172-237-154-236\" cannot list resource \"secrets\" in API group \"\" in the namespace \"calico-system\": no relationship found between node '172-237-154-236' and this object" logger="UnhandledError" Nov 24 00:28:05.401882 kubelet[2709]: W1124 00:28:05.400300 2709 reflector.go:569] object-"calico-system"/"whisker-ca-bundle": failed to list *v1.ConfigMap: configmaps "whisker-ca-bundle" is forbidden: User "system:node:172-237-154-236" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node '172-237-154-236' and this object Nov 24 00:28:05.401882 kubelet[2709]: E1124 00:28:05.401882 2709 reflector.go:166] "Unhandled Error" err="object-\"calico-system\"/\"whisker-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"whisker-ca-bundle\" is forbidden: User \"system:node:172-237-154-236\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"calico-system\": no relationship found between node '172-237-154-236' and this object" logger="UnhandledError" Nov 24 00:28:05.402537 systemd[1]: Created slice kubepods-burstable-pod2fada79a_43af_4af3_bfd2_ddde7f2c0d48.slice - libcontainer container kubepods-burstable-pod2fada79a_43af_4af3_bfd2_ddde7f2c0d48.slice. Nov 24 00:28:05.403761 kubelet[2709]: I1124 00:28:05.400358 2709 status_manager.go:890] "Failed to get status for pod" podUID="2fada79a-43af-4af3-bfd2-ddde7f2c0d48" pod="kube-system/coredns-668d6bf9bc-l5m2s" err="pods \"coredns-668d6bf9bc-l5m2s\" is forbidden: User \"system:node:172-237-154-236\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node '172-237-154-236' and this object" Nov 24 00:28:05.417847 systemd[1]: Created slice kubepods-besteffort-pod37c8f52a_80b1_4ff6_9cad_c82ac32c93d6.slice - libcontainer container kubepods-besteffort-pod37c8f52a_80b1_4ff6_9cad_c82ac32c93d6.slice. Nov 24 00:28:05.430157 kubelet[2709]: I1124 00:28:05.430131 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcjvm\" (UniqueName: \"kubernetes.io/projected/2fada79a-43af-4af3-bfd2-ddde7f2c0d48-kube-api-access-wcjvm\") pod \"coredns-668d6bf9bc-l5m2s\" (UID: \"2fada79a-43af-4af3-bfd2-ddde7f2c0d48\") " pod="kube-system/coredns-668d6bf9bc-l5m2s" Nov 24 00:28:05.430157 kubelet[2709]: I1124 00:28:05.430161 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjv7z\" (UniqueName: \"kubernetes.io/projected/58c2ada9-58c2-46b7-8f97-99d17c62d364-kube-api-access-fjv7z\") pod \"coredns-668d6bf9bc-hj4d7\" (UID: \"58c2ada9-58c2-46b7-8f97-99d17c62d364\") " pod="kube-system/coredns-668d6bf9bc-hj4d7" Nov 24 00:28:05.430268 kubelet[2709]: I1124 00:28:05.430177 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7rvg\" (UniqueName: \"kubernetes.io/projected/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-kube-api-access-g7rvg\") pod \"whisker-7f577c94c7-95j8b\" (UID: \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\") " pod="calico-system/whisker-7f577c94c7-95j8b" Nov 24 00:28:05.430268 kubelet[2709]: I1124 00:28:05.430194 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/96698920-d098-4174-b103-f535ce1caea1-calico-apiserver-certs\") pod \"calico-apiserver-7595d974c5-lnnqz\" (UID: \"96698920-d098-4174-b103-f535ce1caea1\") " pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" Nov 24 00:28:05.430268 kubelet[2709]: I1124 00:28:05.430211 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-ca-bundle\") pod \"whisker-7f577c94c7-95j8b\" (UID: \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\") " pod="calico-system/whisker-7f577c94c7-95j8b" Nov 24 00:28:05.430268 kubelet[2709]: I1124 00:28:05.430249 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58c2ada9-58c2-46b7-8f97-99d17c62d364-config-volume\") pod \"coredns-668d6bf9bc-hj4d7\" (UID: \"58c2ada9-58c2-46b7-8f97-99d17c62d364\") " pod="kube-system/coredns-668d6bf9bc-hj4d7" Nov 24 00:28:05.430268 kubelet[2709]: I1124 00:28:05.430265 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be91967a-4242-4c33-899c-c3fa81f6e4dc-config\") pod \"goldmane-666569f655-g9zxr\" (UID: \"be91967a-4242-4c33-899c-c3fa81f6e4dc\") " pod="calico-system/goldmane-666569f655-g9zxr" Nov 24 00:28:05.430637 kubelet[2709]: I1124 00:28:05.430280 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7208ef7a-cd09-4ec2-a692-f04c8a143f91-tigera-ca-bundle\") pod \"calico-kube-controllers-5b8f657c8-qdb6r\" (UID: \"7208ef7a-cd09-4ec2-a692-f04c8a143f91\") " pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" Nov 24 00:28:05.430637 kubelet[2709]: I1124 00:28:05.430295 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vrxt\" (UniqueName: \"kubernetes.io/projected/d5afe1af-07b4-4767-ad04-88dc81c2e5e8-kube-api-access-9vrxt\") pod \"calico-apiserver-844c564bfd-cldcl\" (UID: \"d5afe1af-07b4-4767-ad04-88dc81c2e5e8\") " pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" Nov 24 00:28:05.430637 kubelet[2709]: I1124 00:28:05.430310 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-backend-key-pair\") pod \"whisker-7f577c94c7-95j8b\" (UID: \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\") " pod="calico-system/whisker-7f577c94c7-95j8b" Nov 24 00:28:05.430637 kubelet[2709]: I1124 00:28:05.430325 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69sxc\" (UniqueName: \"kubernetes.io/projected/96698920-d098-4174-b103-f535ce1caea1-kube-api-access-69sxc\") pod \"calico-apiserver-7595d974c5-lnnqz\" (UID: \"96698920-d098-4174-b103-f535ce1caea1\") " pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" Nov 24 00:28:05.430637 kubelet[2709]: I1124 00:28:05.430345 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/c6030a3d-68f0-4144-8d41-801a65d0dbd4-calico-apiserver-certs\") pod \"calico-apiserver-844c564bfd-tnmdj\" (UID: \"c6030a3d-68f0-4144-8d41-801a65d0dbd4\") " pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" Nov 24 00:28:05.431002 kubelet[2709]: I1124 00:28:05.430360 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7tpq\" (UniqueName: \"kubernetes.io/projected/c6030a3d-68f0-4144-8d41-801a65d0dbd4-kube-api-access-h7tpq\") pod \"calico-apiserver-844c564bfd-tnmdj\" (UID: \"c6030a3d-68f0-4144-8d41-801a65d0dbd4\") " pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" Nov 24 00:28:05.431176 kubelet[2709]: I1124 00:28:05.431054 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dwz2\" (UniqueName: \"kubernetes.io/projected/7208ef7a-cd09-4ec2-a692-f04c8a143f91-kube-api-access-5dwz2\") pod \"calico-kube-controllers-5b8f657c8-qdb6r\" (UID: \"7208ef7a-cd09-4ec2-a692-f04c8a143f91\") " pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" Nov 24 00:28:05.431176 kubelet[2709]: I1124 00:28:05.431112 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/be91967a-4242-4c33-899c-c3fa81f6e4dc-goldmane-key-pair\") pod \"goldmane-666569f655-g9zxr\" (UID: \"be91967a-4242-4c33-899c-c3fa81f6e4dc\") " pod="calico-system/goldmane-666569f655-g9zxr" Nov 24 00:28:05.431176 kubelet[2709]: I1124 00:28:05.431137 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be91967a-4242-4c33-899c-c3fa81f6e4dc-goldmane-ca-bundle\") pod \"goldmane-666569f655-g9zxr\" (UID: \"be91967a-4242-4c33-899c-c3fa81f6e4dc\") " pod="calico-system/goldmane-666569f655-g9zxr" Nov 24 00:28:05.431278 kubelet[2709]: I1124 00:28:05.431266 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2fada79a-43af-4af3-bfd2-ddde7f2c0d48-config-volume\") pod \"coredns-668d6bf9bc-l5m2s\" (UID: \"2fada79a-43af-4af3-bfd2-ddde7f2c0d48\") " pod="kube-system/coredns-668d6bf9bc-l5m2s" Nov 24 00:28:05.431305 kubelet[2709]: I1124 00:28:05.431286 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/d5afe1af-07b4-4767-ad04-88dc81c2e5e8-calico-apiserver-certs\") pod \"calico-apiserver-844c564bfd-cldcl\" (UID: \"d5afe1af-07b4-4767-ad04-88dc81c2e5e8\") " pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" Nov 24 00:28:05.431305 kubelet[2709]: I1124 00:28:05.431303 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m87kj\" (UniqueName: \"kubernetes.io/projected/be91967a-4242-4c33-899c-c3fa81f6e4dc-kube-api-access-m87kj\") pod \"goldmane-666569f655-g9zxr\" (UID: \"be91967a-4242-4c33-899c-c3fa81f6e4dc\") " pod="calico-system/goldmane-666569f655-g9zxr" Nov 24 00:28:05.431712 systemd[1]: Created slice kubepods-besteffort-podc6030a3d_68f0_4144_8d41_801a65d0dbd4.slice - libcontainer container kubepods-besteffort-podc6030a3d_68f0_4144_8d41_801a65d0dbd4.slice. Nov 24 00:28:05.443529 systemd[1]: Created slice kubepods-besteffort-pod7208ef7a_cd09_4ec2_a692_f04c8a143f91.slice - libcontainer container kubepods-besteffort-pod7208ef7a_cd09_4ec2_a692_f04c8a143f91.slice. Nov 24 00:28:05.453653 systemd[1]: Created slice kubepods-burstable-pod58c2ada9_58c2_46b7_8f97_99d17c62d364.slice - libcontainer container kubepods-burstable-pod58c2ada9_58c2_46b7_8f97_99d17c62d364.slice. Nov 24 00:28:05.463053 systemd[1]: Created slice kubepods-besteffort-podd5afe1af_07b4_4767_ad04_88dc81c2e5e8.slice - libcontainer container kubepods-besteffort-podd5afe1af_07b4_4767_ad04_88dc81c2e5e8.slice. Nov 24 00:28:05.473275 systemd[1]: Created slice kubepods-besteffort-podbe91967a_4242_4c33_899c_c3fa81f6e4dc.slice - libcontainer container kubepods-besteffort-podbe91967a_4242_4c33_899c_c3fa81f6e4dc.slice. Nov 24 00:28:05.481234 systemd[1]: Created slice kubepods-besteffort-pod96698920_d098_4174_b103_f535ce1caea1.slice - libcontainer container kubepods-besteffort-pod96698920_d098_4174_b103_f535ce1caea1.slice. Nov 24 00:28:05.712162 kubelet[2709]: E1124 00:28:05.712129 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:05.712849 containerd[1559]: time="2025-11-24T00:28:05.712793305Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-l5m2s,Uid:2fada79a-43af-4af3-bfd2-ddde7f2c0d48,Namespace:kube-system,Attempt:0,}" Nov 24 00:28:05.743315 containerd[1559]: time="2025-11-24T00:28:05.743272010Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-844c564bfd-tnmdj,Uid:c6030a3d-68f0-4144-8d41-801a65d0dbd4,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:28:05.750283 containerd[1559]: time="2025-11-24T00:28:05.749642394Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5b8f657c8-qdb6r,Uid:7208ef7a-cd09-4ec2-a692-f04c8a143f91,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:05.762337 kubelet[2709]: E1124 00:28:05.762316 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:05.765869 containerd[1559]: time="2025-11-24T00:28:05.765631692Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hj4d7,Uid:58c2ada9-58c2-46b7-8f97-99d17c62d364,Namespace:kube-system,Attempt:0,}" Nov 24 00:28:05.767870 systemd[1]: Created slice kubepods-besteffort-podd80649f2_a0ab_46f8_9106_ed888d934bff.slice - libcontainer container kubepods-besteffort-podd80649f2_a0ab_46f8_9106_ed888d934bff.slice. Nov 24 00:28:05.770938 containerd[1559]: time="2025-11-24T00:28:05.769708814Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-844c564bfd-cldcl,Uid:d5afe1af-07b4-4767-ad04-88dc81c2e5e8,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:28:05.775177 containerd[1559]: time="2025-11-24T00:28:05.775138406Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5vjmh,Uid:d80649f2-a0ab-46f8-9106-ed888d934bff,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:05.777690 containerd[1559]: time="2025-11-24T00:28:05.777670218Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-g9zxr,Uid:be91967a-4242-4c33-899c-c3fa81f6e4dc,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:05.798251 containerd[1559]: time="2025-11-24T00:28:05.797868148Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7595d974c5-lnnqz,Uid:96698920-d098-4174-b103-f535ce1caea1,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:28:05.872391 containerd[1559]: time="2025-11-24T00:28:05.872351285Z" level=error msg="Failed to destroy network for sandbox \"0b9d1c46f1180b0cf8a48c83259a2e6472a39b247bebcf838d57cd36d65c0801\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.880611 containerd[1559]: time="2025-11-24T00:28:05.880583139Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-l5m2s,Uid:2fada79a-43af-4af3-bfd2-ddde7f2c0d48,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b9d1c46f1180b0cf8a48c83259a2e6472a39b247bebcf838d57cd36d65c0801\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.881116 kubelet[2709]: E1124 00:28:05.881057 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b9d1c46f1180b0cf8a48c83259a2e6472a39b247bebcf838d57cd36d65c0801\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.881694 kubelet[2709]: E1124 00:28:05.881654 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b9d1c46f1180b0cf8a48c83259a2e6472a39b247bebcf838d57cd36d65c0801\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-l5m2s" Nov 24 00:28:05.881872 kubelet[2709]: E1124 00:28:05.881831 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b9d1c46f1180b0cf8a48c83259a2e6472a39b247bebcf838d57cd36d65c0801\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-l5m2s" Nov 24 00:28:05.882166 kubelet[2709]: E1124 00:28:05.882005 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-l5m2s_kube-system(2fada79a-43af-4af3-bfd2-ddde7f2c0d48)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-l5m2s_kube-system(2fada79a-43af-4af3-bfd2-ddde7f2c0d48)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0b9d1c46f1180b0cf8a48c83259a2e6472a39b247bebcf838d57cd36d65c0801\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-l5m2s" podUID="2fada79a-43af-4af3-bfd2-ddde7f2c0d48" Nov 24 00:28:05.887942 kubelet[2709]: E1124 00:28:05.887773 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:05.894133 containerd[1559]: time="2025-11-24T00:28:05.894077616Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Nov 24 00:28:05.940609 containerd[1559]: time="2025-11-24T00:28:05.940487329Z" level=error msg="Failed to destroy network for sandbox \"6aacb28e6a984fbd0959e18e28050a0296211de921cf26d0ca2afb2fda6eb58d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.943787 containerd[1559]: time="2025-11-24T00:28:05.943750281Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7595d974c5-lnnqz,Uid:96698920-d098-4174-b103-f535ce1caea1,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"6aacb28e6a984fbd0959e18e28050a0296211de921cf26d0ca2afb2fda6eb58d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.944397 kubelet[2709]: E1124 00:28:05.944348 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6aacb28e6a984fbd0959e18e28050a0296211de921cf26d0ca2afb2fda6eb58d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.944456 kubelet[2709]: E1124 00:28:05.944403 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6aacb28e6a984fbd0959e18e28050a0296211de921cf26d0ca2afb2fda6eb58d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" Nov 24 00:28:05.944456 kubelet[2709]: E1124 00:28:05.944423 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6aacb28e6a984fbd0959e18e28050a0296211de921cf26d0ca2afb2fda6eb58d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" Nov 24 00:28:05.944658 kubelet[2709]: E1124 00:28:05.944624 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7595d974c5-lnnqz_calico-apiserver(96698920-d098-4174-b103-f535ce1caea1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7595d974c5-lnnqz_calico-apiserver(96698920-d098-4174-b103-f535ce1caea1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6aacb28e6a984fbd0959e18e28050a0296211de921cf26d0ca2afb2fda6eb58d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:28:05.990033 containerd[1559]: time="2025-11-24T00:28:05.989869134Z" level=error msg="Failed to destroy network for sandbox \"bbe7cca477932a8f7a5803e46b8952dbf317e3995e308747428ad7d21b7fbaf0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.993042 containerd[1559]: time="2025-11-24T00:28:05.993002415Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-844c564bfd-tnmdj,Uid:c6030a3d-68f0-4144-8d41-801a65d0dbd4,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bbe7cca477932a8f7a5803e46b8952dbf317e3995e308747428ad7d21b7fbaf0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.993571 kubelet[2709]: E1124 00:28:05.993448 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bbe7cca477932a8f7a5803e46b8952dbf317e3995e308747428ad7d21b7fbaf0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:05.993689 kubelet[2709]: E1124 00:28:05.993667 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bbe7cca477932a8f7a5803e46b8952dbf317e3995e308747428ad7d21b7fbaf0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" Nov 24 00:28:05.993929 kubelet[2709]: E1124 00:28:05.993787 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bbe7cca477932a8f7a5803e46b8952dbf317e3995e308747428ad7d21b7fbaf0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" Nov 24 00:28:05.994104 kubelet[2709]: E1124 00:28:05.993898 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-844c564bfd-tnmdj_calico-apiserver(c6030a3d-68f0-4144-8d41-801a65d0dbd4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-844c564bfd-tnmdj_calico-apiserver(c6030a3d-68f0-4144-8d41-801a65d0dbd4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bbe7cca477932a8f7a5803e46b8952dbf317e3995e308747428ad7d21b7fbaf0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:28:06.020377 containerd[1559]: time="2025-11-24T00:28:06.020307689Z" level=error msg="Failed to destroy network for sandbox \"7451bdfe27578a93dd2e8a0ce75287cf9f844e58f73d1b6b529d251cc85110cd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.022712 containerd[1559]: time="2025-11-24T00:28:06.022424050Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-844c564bfd-cldcl,Uid:d5afe1af-07b4-4767-ad04-88dc81c2e5e8,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7451bdfe27578a93dd2e8a0ce75287cf9f844e58f73d1b6b529d251cc85110cd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.023855 kubelet[2709]: E1124 00:28:06.022945 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7451bdfe27578a93dd2e8a0ce75287cf9f844e58f73d1b6b529d251cc85110cd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.023855 kubelet[2709]: E1124 00:28:06.022999 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7451bdfe27578a93dd2e8a0ce75287cf9f844e58f73d1b6b529d251cc85110cd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" Nov 24 00:28:06.023855 kubelet[2709]: E1124 00:28:06.023022 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7451bdfe27578a93dd2e8a0ce75287cf9f844e58f73d1b6b529d251cc85110cd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" Nov 24 00:28:06.024542 kubelet[2709]: E1124 00:28:06.023059 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-844c564bfd-cldcl_calico-apiserver(d5afe1af-07b4-4767-ad04-88dc81c2e5e8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-844c564bfd-cldcl_calico-apiserver(d5afe1af-07b4-4767-ad04-88dc81c2e5e8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7451bdfe27578a93dd2e8a0ce75287cf9f844e58f73d1b6b529d251cc85110cd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:28:06.044973 containerd[1559]: time="2025-11-24T00:28:06.044949461Z" level=error msg="Failed to destroy network for sandbox \"432278399252ba241a24d4f541764cf2a1933b0cd5c4d70e0e49bb9eee91ac17\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.046296 containerd[1559]: time="2025-11-24T00:28:06.046271902Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5vjmh,Uid:d80649f2-a0ab-46f8-9106-ed888d934bff,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"432278399252ba241a24d4f541764cf2a1933b0cd5c4d70e0e49bb9eee91ac17\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.046540 kubelet[2709]: E1124 00:28:06.046518 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"432278399252ba241a24d4f541764cf2a1933b0cd5c4d70e0e49bb9eee91ac17\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.046729 kubelet[2709]: E1124 00:28:06.046713 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"432278399252ba241a24d4f541764cf2a1933b0cd5c4d70e0e49bb9eee91ac17\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5vjmh" Nov 24 00:28:06.046804 kubelet[2709]: E1124 00:28:06.046787 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"432278399252ba241a24d4f541764cf2a1933b0cd5c4d70e0e49bb9eee91ac17\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5vjmh" Nov 24 00:28:06.047388 kubelet[2709]: E1124 00:28:06.047345 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"432278399252ba241a24d4f541764cf2a1933b0cd5c4d70e0e49bb9eee91ac17\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:06.047794 containerd[1559]: time="2025-11-24T00:28:06.047758772Z" level=error msg="Failed to destroy network for sandbox \"f7ae0b5c4e4c1600b6c0804b98fe64cfa56d3615fdc19944c61d92e3008a399d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.048785 containerd[1559]: time="2025-11-24T00:28:06.048751833Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-g9zxr,Uid:be91967a-4242-4c33-899c-c3fa81f6e4dc,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7ae0b5c4e4c1600b6c0804b98fe64cfa56d3615fdc19944c61d92e3008a399d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.049011 kubelet[2709]: E1124 00:28:06.048903 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7ae0b5c4e4c1600b6c0804b98fe64cfa56d3615fdc19944c61d92e3008a399d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.049011 kubelet[2709]: E1124 00:28:06.049015 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7ae0b5c4e4c1600b6c0804b98fe64cfa56d3615fdc19944c61d92e3008a399d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-g9zxr" Nov 24 00:28:06.049011 kubelet[2709]: E1124 00:28:06.049034 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7ae0b5c4e4c1600b6c0804b98fe64cfa56d3615fdc19944c61d92e3008a399d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-g9zxr" Nov 24 00:28:06.049539 kubelet[2709]: E1124 00:28:06.049068 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-g9zxr_calico-system(be91967a-4242-4c33-899c-c3fa81f6e4dc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-g9zxr_calico-system(be91967a-4242-4c33-899c-c3fa81f6e4dc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f7ae0b5c4e4c1600b6c0804b98fe64cfa56d3615fdc19944c61d92e3008a399d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:28:06.049900 containerd[1559]: time="2025-11-24T00:28:06.049745723Z" level=error msg="Failed to destroy network for sandbox \"4c945e0b0f7f491164d6d30bce3cfbfe3886b0b558ee3431a5ab700e6f6c96a3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.052710 containerd[1559]: time="2025-11-24T00:28:06.052663625Z" level=error msg="Failed to destroy network for sandbox \"eb23bfc3eb496f39b553699e5ef6c4d7f23891041342f31803aec30ce130de0c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.053056 containerd[1559]: time="2025-11-24T00:28:06.052985705Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5b8f657c8-qdb6r,Uid:7208ef7a-cd09-4ec2-a692-f04c8a143f91,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4c945e0b0f7f491164d6d30bce3cfbfe3886b0b558ee3431a5ab700e6f6c96a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.053254 kubelet[2709]: E1124 00:28:06.053225 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4c945e0b0f7f491164d6d30bce3cfbfe3886b0b558ee3431a5ab700e6f6c96a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.053323 kubelet[2709]: E1124 00:28:06.053258 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4c945e0b0f7f491164d6d30bce3cfbfe3886b0b558ee3431a5ab700e6f6c96a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" Nov 24 00:28:06.053323 kubelet[2709]: E1124 00:28:06.053272 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4c945e0b0f7f491164d6d30bce3cfbfe3886b0b558ee3431a5ab700e6f6c96a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" Nov 24 00:28:06.053729 kubelet[2709]: E1124 00:28:06.053298 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5b8f657c8-qdb6r_calico-system(7208ef7a-cd09-4ec2-a692-f04c8a143f91)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5b8f657c8-qdb6r_calico-system(7208ef7a-cd09-4ec2-a692-f04c8a143f91)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4c945e0b0f7f491164d6d30bce3cfbfe3886b0b558ee3431a5ab700e6f6c96a3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:28:06.053869 containerd[1559]: time="2025-11-24T00:28:06.053793666Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hj4d7,Uid:58c2ada9-58c2-46b7-8f97-99d17c62d364,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb23bfc3eb496f39b553699e5ef6c4d7f23891041342f31803aec30ce130de0c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.054110 kubelet[2709]: E1124 00:28:06.053903 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb23bfc3eb496f39b553699e5ef6c4d7f23891041342f31803aec30ce130de0c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:06.054110 kubelet[2709]: E1124 00:28:06.054077 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb23bfc3eb496f39b553699e5ef6c4d7f23891041342f31803aec30ce130de0c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-hj4d7" Nov 24 00:28:06.054110 kubelet[2709]: E1124 00:28:06.054089 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb23bfc3eb496f39b553699e5ef6c4d7f23891041342f31803aec30ce130de0c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-hj4d7" Nov 24 00:28:06.054214 kubelet[2709]: E1124 00:28:06.054147 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-hj4d7_kube-system(58c2ada9-58c2-46b7-8f97-99d17c62d364)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-hj4d7_kube-system(58c2ada9-58c2-46b7-8f97-99d17c62d364)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"eb23bfc3eb496f39b553699e5ef6c4d7f23891041342f31803aec30ce130de0c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-hj4d7" podUID="58c2ada9-58c2-46b7-8f97-99d17c62d364" Nov 24 00:28:06.536944 kubelet[2709]: E1124 00:28:06.536630 2709 secret.go:189] Couldn't get secret calico-system/whisker-backend-key-pair: failed to sync secret cache: timed out waiting for the condition Nov 24 00:28:06.536944 kubelet[2709]: E1124 00:28:06.536731 2709 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-backend-key-pair podName:37c8f52a-80b1-4ff6-9cad-c82ac32c93d6 nodeName:}" failed. No retries permitted until 2025-11-24 00:28:07.036708367 +0000 UTC m=+26.372516969 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "whisker-backend-key-pair" (UniqueName: "kubernetes.io/secret/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-backend-key-pair") pod "whisker-7f577c94c7-95j8b" (UID: "37c8f52a-80b1-4ff6-9cad-c82ac32c93d6") : failed to sync secret cache: timed out waiting for the condition Nov 24 00:28:06.648906 systemd[1]: run-netns-cni\x2dc96b6664\x2d42de\x2d6245\x2daa8c\x2d2fd94f49b20a.mount: Deactivated successfully. Nov 24 00:28:06.650040 systemd[1]: run-netns-cni\x2dd2ca8668\x2d5a25\x2d8ef8\x2d8b5d\x2d470e91baffd7.mount: Deactivated successfully. Nov 24 00:28:06.650117 systemd[1]: run-netns-cni\x2d806610ee\x2da78b\x2d9319\x2d5062\x2d22b522b8d203.mount: Deactivated successfully. Nov 24 00:28:06.650248 systemd[1]: run-netns-cni\x2da13944b2\x2dc7a5\x2d80ce\x2d898a\x2d4917bc1f8b9d.mount: Deactivated successfully. Nov 24 00:28:07.237652 containerd[1559]: time="2025-11-24T00:28:07.237409837Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7f577c94c7-95j8b,Uid:37c8f52a-80b1-4ff6-9cad-c82ac32c93d6,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:07.319148 containerd[1559]: time="2025-11-24T00:28:07.319083368Z" level=error msg="Failed to destroy network for sandbox \"3efc459a9b1eb930c60d57e81fb06d64bbde9ccb736131742c617182032e0aa6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:07.322130 systemd[1]: run-netns-cni\x2d3dc14992\x2d4c34\x2d5755\x2dea2c\x2d254691455498.mount: Deactivated successfully. Nov 24 00:28:07.324082 containerd[1559]: time="2025-11-24T00:28:07.323970430Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7f577c94c7-95j8b,Uid:37c8f52a-80b1-4ff6-9cad-c82ac32c93d6,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3efc459a9b1eb930c60d57e81fb06d64bbde9ccb736131742c617182032e0aa6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:07.324396 kubelet[2709]: E1124 00:28:07.324249 2709 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3efc459a9b1eb930c60d57e81fb06d64bbde9ccb736131742c617182032e0aa6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:28:07.325515 kubelet[2709]: E1124 00:28:07.324660 2709 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3efc459a9b1eb930c60d57e81fb06d64bbde9ccb736131742c617182032e0aa6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7f577c94c7-95j8b" Nov 24 00:28:07.325515 kubelet[2709]: E1124 00:28:07.324689 2709 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3efc459a9b1eb930c60d57e81fb06d64bbde9ccb736131742c617182032e0aa6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7f577c94c7-95j8b" Nov 24 00:28:07.325515 kubelet[2709]: E1124 00:28:07.325135 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-7f577c94c7-95j8b_calico-system(37c8f52a-80b1-4ff6-9cad-c82ac32c93d6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-7f577c94c7-95j8b_calico-system(37c8f52a-80b1-4ff6-9cad-c82ac32c93d6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3efc459a9b1eb930c60d57e81fb06d64bbde9ccb736131742c617182032e0aa6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7f577c94c7-95j8b" podUID="37c8f52a-80b1-4ff6-9cad-c82ac32c93d6" Nov 24 00:28:08.523556 kubelet[2709]: I1124 00:28:08.523511 2709 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 00:28:08.523996 kubelet[2709]: E1124 00:28:08.523809 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:08.893110 kubelet[2709]: E1124 00:28:08.892905 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:09.428348 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2364223704.mount: Deactivated successfully. Nov 24 00:28:09.450565 containerd[1559]: time="2025-11-24T00:28:09.450523203Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:09.451304 containerd[1559]: time="2025-11-24T00:28:09.451167263Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156883675" Nov 24 00:28:09.451817 containerd[1559]: time="2025-11-24T00:28:09.451787043Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:09.453213 containerd[1559]: time="2025-11-24T00:28:09.453187484Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:28:09.453702 containerd[1559]: time="2025-11-24T00:28:09.453669334Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 3.559550368s" Nov 24 00:28:09.453746 containerd[1559]: time="2025-11-24T00:28:09.453704924Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Nov 24 00:28:09.469517 containerd[1559]: time="2025-11-24T00:28:09.468814742Z" level=info msg="CreateContainer within sandbox \"89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Nov 24 00:28:09.484327 containerd[1559]: time="2025-11-24T00:28:09.484081699Z" level=info msg="Container 3ac0694ad5c942aa41be8f725ef1ab998ffc011e01a67afebf780164dfabe655: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:28:09.488045 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1264216032.mount: Deactivated successfully. Nov 24 00:28:09.493753 containerd[1559]: time="2025-11-24T00:28:09.493717434Z" level=info msg="CreateContainer within sandbox \"89a9904789873405394bed106d50cd126af4cf4b76028a1a2ec6a5a1730fb673\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"3ac0694ad5c942aa41be8f725ef1ab998ffc011e01a67afebf780164dfabe655\"" Nov 24 00:28:09.495269 containerd[1559]: time="2025-11-24T00:28:09.495240655Z" level=info msg="StartContainer for \"3ac0694ad5c942aa41be8f725ef1ab998ffc011e01a67afebf780164dfabe655\"" Nov 24 00:28:09.496717 containerd[1559]: time="2025-11-24T00:28:09.496683156Z" level=info msg="connecting to shim 3ac0694ad5c942aa41be8f725ef1ab998ffc011e01a67afebf780164dfabe655" address="unix:///run/containerd/s/dc465aa4d2a3f40975a175965f23c6e15aef3c4794f2fa167bcc6014e7cd1389" protocol=ttrpc version=3 Nov 24 00:28:09.537045 systemd[1]: Started cri-containerd-3ac0694ad5c942aa41be8f725ef1ab998ffc011e01a67afebf780164dfabe655.scope - libcontainer container 3ac0694ad5c942aa41be8f725ef1ab998ffc011e01a67afebf780164dfabe655. Nov 24 00:28:09.608847 containerd[1559]: time="2025-11-24T00:28:09.608643302Z" level=info msg="StartContainer for \"3ac0694ad5c942aa41be8f725ef1ab998ffc011e01a67afebf780164dfabe655\" returns successfully" Nov 24 00:28:09.686164 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Nov 24 00:28:09.686256 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Nov 24 00:28:09.865604 kubelet[2709]: I1124 00:28:09.865073 2709 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7rvg\" (UniqueName: \"kubernetes.io/projected/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-kube-api-access-g7rvg\") pod \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\" (UID: \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\") " Nov 24 00:28:09.865604 kubelet[2709]: I1124 00:28:09.865122 2709 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-ca-bundle\") pod \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\" (UID: \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\") " Nov 24 00:28:09.865604 kubelet[2709]: I1124 00:28:09.865143 2709 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-backend-key-pair\") pod \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\" (UID: \"37c8f52a-80b1-4ff6-9cad-c82ac32c93d6\") " Nov 24 00:28:09.869515 kubelet[2709]: I1124 00:28:09.869489 2709 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "37c8f52a-80b1-4ff6-9cad-c82ac32c93d6" (UID: "37c8f52a-80b1-4ff6-9cad-c82ac32c93d6"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Nov 24 00:28:09.869975 kubelet[2709]: I1124 00:28:09.869938 2709 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "37c8f52a-80b1-4ff6-9cad-c82ac32c93d6" (UID: "37c8f52a-80b1-4ff6-9cad-c82ac32c93d6"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Nov 24 00:28:09.873156 kubelet[2709]: I1124 00:28:09.873120 2709 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-kube-api-access-g7rvg" (OuterVolumeSpecName: "kube-api-access-g7rvg") pod "37c8f52a-80b1-4ff6-9cad-c82ac32c93d6" (UID: "37c8f52a-80b1-4ff6-9cad-c82ac32c93d6"). InnerVolumeSpecName "kube-api-access-g7rvg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Nov 24 00:28:09.903551 kubelet[2709]: E1124 00:28:09.903518 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:09.924282 systemd[1]: Removed slice kubepods-besteffort-pod37c8f52a_80b1_4ff6_9cad_c82ac32c93d6.slice - libcontainer container kubepods-besteffort-pod37c8f52a_80b1_4ff6_9cad_c82ac32c93d6.slice. Nov 24 00:28:09.938951 kubelet[2709]: I1124 00:28:09.937981 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-pvs72" podStartSLOduration=1.088363004 podStartE2EDuration="9.937966816s" podCreationTimestamp="2025-11-24 00:28:00 +0000 UTC" firstStartedPulling="2025-11-24 00:28:00.604997373 +0000 UTC m=+19.940805975" lastFinishedPulling="2025-11-24 00:28:09.454601185 +0000 UTC m=+28.790409787" observedRunningTime="2025-11-24 00:28:09.934102964 +0000 UTC m=+29.269911566" watchObservedRunningTime="2025-11-24 00:28:09.937966816 +0000 UTC m=+29.273775438" Nov 24 00:28:09.965940 kubelet[2709]: I1124 00:28:09.965674 2709 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-g7rvg\" (UniqueName: \"kubernetes.io/projected/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-kube-api-access-g7rvg\") on node \"172-237-154-236\" DevicePath \"\"" Nov 24 00:28:09.967687 kubelet[2709]: I1124 00:28:09.967667 2709 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-backend-key-pair\") on node \"172-237-154-236\" DevicePath \"\"" Nov 24 00:28:09.967793 kubelet[2709]: I1124 00:28:09.967782 2709 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6-whisker-ca-bundle\") on node \"172-237-154-236\" DevicePath \"\"" Nov 24 00:28:10.007868 systemd[1]: Created slice kubepods-besteffort-pod64ccd4c8_759d_46f3_b211_0873db45cfc8.slice - libcontainer container kubepods-besteffort-pod64ccd4c8_759d_46f3_b211_0873db45cfc8.slice. Nov 24 00:28:10.068312 kubelet[2709]: I1124 00:28:10.068279 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/64ccd4c8-759d-46f3-b211-0873db45cfc8-whisker-backend-key-pair\") pod \"whisker-67b5f8fb49-hhhrv\" (UID: \"64ccd4c8-759d-46f3-b211-0873db45cfc8\") " pod="calico-system/whisker-67b5f8fb49-hhhrv" Nov 24 00:28:10.069060 kubelet[2709]: I1124 00:28:10.069005 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64ccd4c8-759d-46f3-b211-0873db45cfc8-whisker-ca-bundle\") pod \"whisker-67b5f8fb49-hhhrv\" (UID: \"64ccd4c8-759d-46f3-b211-0873db45cfc8\") " pod="calico-system/whisker-67b5f8fb49-hhhrv" Nov 24 00:28:10.069199 kubelet[2709]: I1124 00:28:10.069148 2709 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2q9n\" (UniqueName: \"kubernetes.io/projected/64ccd4c8-759d-46f3-b211-0873db45cfc8-kube-api-access-m2q9n\") pod \"whisker-67b5f8fb49-hhhrv\" (UID: \"64ccd4c8-759d-46f3-b211-0873db45cfc8\") " pod="calico-system/whisker-67b5f8fb49-hhhrv" Nov 24 00:28:10.316610 containerd[1559]: time="2025-11-24T00:28:10.316554165Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-67b5f8fb49-hhhrv,Uid:64ccd4c8-759d-46f3-b211-0873db45cfc8,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:10.437324 systemd[1]: var-lib-kubelet-pods-37c8f52a\x2d80b1\x2d4ff6\x2d9cad\x2dc82ac32c93d6-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Nov 24 00:28:10.437464 systemd[1]: var-lib-kubelet-pods-37c8f52a\x2d80b1\x2d4ff6\x2d9cad\x2dc82ac32c93d6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dg7rvg.mount: Deactivated successfully. Nov 24 00:28:10.491203 systemd-networkd[1439]: cali9c92fb7616a: Link UP Nov 24 00:28:10.493266 systemd-networkd[1439]: cali9c92fb7616a: Gained carrier Nov 24 00:28:10.516998 containerd[1559]: 2025-11-24 00:28:10.345 [INFO][3844] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Nov 24 00:28:10.516998 containerd[1559]: 2025-11-24 00:28:10.384 [INFO][3844] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0 whisker-67b5f8fb49- calico-system 64ccd4c8-759d-46f3-b211-0873db45cfc8 904 0 2025-11-24 00:28:09 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:67b5f8fb49 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s 172-237-154-236 whisker-67b5f8fb49-hhhrv eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali9c92fb7616a [] [] }} ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Namespace="calico-system" Pod="whisker-67b5f8fb49-hhhrv" WorkloadEndpoint="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-" Nov 24 00:28:10.516998 containerd[1559]: 2025-11-24 00:28:10.384 [INFO][3844] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Namespace="calico-system" Pod="whisker-67b5f8fb49-hhhrv" WorkloadEndpoint="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" Nov 24 00:28:10.516998 containerd[1559]: 2025-11-24 00:28:10.411 [INFO][3855] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" HandleID="k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Workload="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.411 [INFO][3855] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" HandleID="k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Workload="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f120), Attrs:map[string]string{"namespace":"calico-system", "node":"172-237-154-236", "pod":"whisker-67b5f8fb49-hhhrv", "timestamp":"2025-11-24 00:28:10.411374733 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.411 [INFO][3855] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.411 [INFO][3855] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.412 [INFO][3855] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.418 [INFO][3855] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" host="172-237-154-236" Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.423 [INFO][3855] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.433 [INFO][3855] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.435 [INFO][3855] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:10.517454 containerd[1559]: 2025-11-24 00:28:10.440 [INFO][3855] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:10.517649 containerd[1559]: 2025-11-24 00:28:10.440 [INFO][3855] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" host="172-237-154-236" Nov 24 00:28:10.517649 containerd[1559]: 2025-11-24 00:28:10.442 [INFO][3855] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59 Nov 24 00:28:10.517649 containerd[1559]: 2025-11-24 00:28:10.453 [INFO][3855] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" host="172-237-154-236" Nov 24 00:28:10.517649 containerd[1559]: 2025-11-24 00:28:10.465 [INFO][3855] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.129/26] block=192.168.40.128/26 handle="k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" host="172-237-154-236" Nov 24 00:28:10.517649 containerd[1559]: 2025-11-24 00:28:10.465 [INFO][3855] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.129/26] handle="k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" host="172-237-154-236" Nov 24 00:28:10.517649 containerd[1559]: 2025-11-24 00:28:10.466 [INFO][3855] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:10.517649 containerd[1559]: 2025-11-24 00:28:10.466 [INFO][3855] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.129/26] IPv6=[] ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" HandleID="k8s-pod-network.5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Workload="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" Nov 24 00:28:10.517843 containerd[1559]: 2025-11-24 00:28:10.476 [INFO][3844] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Namespace="calico-system" Pod="whisker-67b5f8fb49-hhhrv" WorkloadEndpoint="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0", GenerateName:"whisker-67b5f8fb49-", Namespace:"calico-system", SelfLink:"", UID:"64ccd4c8-759d-46f3-b211-0873db45cfc8", ResourceVersion:"904", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 28, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"67b5f8fb49", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"whisker-67b5f8fb49-hhhrv", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.40.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali9c92fb7616a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:10.517843 containerd[1559]: 2025-11-24 00:28:10.476 [INFO][3844] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.129/32] ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Namespace="calico-system" Pod="whisker-67b5f8fb49-hhhrv" WorkloadEndpoint="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" Nov 24 00:28:10.517937 containerd[1559]: 2025-11-24 00:28:10.476 [INFO][3844] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9c92fb7616a ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Namespace="calico-system" Pod="whisker-67b5f8fb49-hhhrv" WorkloadEndpoint="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" Nov 24 00:28:10.517937 containerd[1559]: 2025-11-24 00:28:10.493 [INFO][3844] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Namespace="calico-system" Pod="whisker-67b5f8fb49-hhhrv" WorkloadEndpoint="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" Nov 24 00:28:10.517977 containerd[1559]: 2025-11-24 00:28:10.494 [INFO][3844] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Namespace="calico-system" Pod="whisker-67b5f8fb49-hhhrv" WorkloadEndpoint="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0", GenerateName:"whisker-67b5f8fb49-", Namespace:"calico-system", SelfLink:"", UID:"64ccd4c8-759d-46f3-b211-0873db45cfc8", ResourceVersion:"904", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 28, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"67b5f8fb49", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59", Pod:"whisker-67b5f8fb49-hhhrv", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.40.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali9c92fb7616a", MAC:"ca:ae:de:e8:e0:e1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:10.518031 containerd[1559]: 2025-11-24 00:28:10.512 [INFO][3844] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" Namespace="calico-system" Pod="whisker-67b5f8fb49-hhhrv" WorkloadEndpoint="172--237--154--236-k8s-whisker--67b5f8fb49--hhhrv-eth0" Nov 24 00:28:10.558128 containerd[1559]: time="2025-11-24T00:28:10.558070566Z" level=info msg="connecting to shim 5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59" address="unix:///run/containerd/s/0a9c011316714c8eefb2570d22ee53a9548bfec57295f5b76613b18c32a36077" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:10.586062 systemd[1]: Started cri-containerd-5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59.scope - libcontainer container 5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59. Nov 24 00:28:10.633410 containerd[1559]: time="2025-11-24T00:28:10.633368184Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-67b5f8fb49-hhhrv,Uid:64ccd4c8-759d-46f3-b211-0873db45cfc8,Namespace:calico-system,Attempt:0,} returns sandbox id \"5ec658cb9c2a7baffb9c97c59d8c3edb8447d447d1a6fe0c9f9f2d3bdb853a59\"" Nov 24 00:28:10.635166 containerd[1559]: time="2025-11-24T00:28:10.635133195Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 24 00:28:10.759301 kubelet[2709]: I1124 00:28:10.759241 2709 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37c8f52a-80b1-4ff6-9cad-c82ac32c93d6" path="/var/lib/kubelet/pods/37c8f52a-80b1-4ff6-9cad-c82ac32c93d6/volumes" Nov 24 00:28:10.776836 containerd[1559]: time="2025-11-24T00:28:10.776762005Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:10.778135 containerd[1559]: time="2025-11-24T00:28:10.778054846Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 24 00:28:10.778193 containerd[1559]: time="2025-11-24T00:28:10.778178336Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 24 00:28:10.778375 kubelet[2709]: E1124 00:28:10.778327 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:28:10.778375 kubelet[2709]: E1124 00:28:10.778371 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:28:10.779824 kubelet[2709]: E1124 00:28:10.779775 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:4781a789a9cf40d795f128549fb7a9fc,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2q9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-67b5f8fb49-hhhrv_calico-system(64ccd4c8-759d-46f3-b211-0873db45cfc8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:10.781796 containerd[1559]: time="2025-11-24T00:28:10.781757638Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 24 00:28:10.904923 kubelet[2709]: E1124 00:28:10.904865 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:10.906169 containerd[1559]: time="2025-11-24T00:28:10.906058760Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:10.907362 containerd[1559]: time="2025-11-24T00:28:10.907307081Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 24 00:28:10.907421 containerd[1559]: time="2025-11-24T00:28:10.907384261Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 24 00:28:10.907554 kubelet[2709]: E1124 00:28:10.907528 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:28:10.907596 kubelet[2709]: E1124 00:28:10.907558 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:28:10.907720 kubelet[2709]: E1124 00:28:10.907644 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m2q9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-67b5f8fb49-hhhrv_calico-system(64ccd4c8-759d-46f3-b211-0873db45cfc8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:10.909934 kubelet[2709]: E1124 00:28:10.909704 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:28:11.801777 systemd-networkd[1439]: vxlan.calico: Link UP Nov 24 00:28:11.801786 systemd-networkd[1439]: vxlan.calico: Gained carrier Nov 24 00:28:11.908948 kubelet[2709]: E1124 00:28:11.908477 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:11.911435 kubelet[2709]: E1124 00:28:11.911408 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:28:12.275109 systemd-networkd[1439]: cali9c92fb7616a: Gained IPv6LL Nov 24 00:28:13.619180 systemd-networkd[1439]: vxlan.calico: Gained IPv6LL Nov 24 00:28:16.757806 kubelet[2709]: E1124 00:28:16.757272 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:16.758378 containerd[1559]: time="2025-11-24T00:28:16.757477064Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-l5m2s,Uid:2fada79a-43af-4af3-bfd2-ddde7f2c0d48,Namespace:kube-system,Attempt:0,}" Nov 24 00:28:16.859317 systemd-networkd[1439]: cali610973304d4: Link UP Nov 24 00:28:16.859536 systemd-networkd[1439]: cali610973304d4: Gained carrier Nov 24 00:28:16.874240 containerd[1559]: 2025-11-24 00:28:16.793 [INFO][4165] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0 coredns-668d6bf9bc- kube-system 2fada79a-43af-4af3-bfd2-ddde7f2c0d48 823 0 2025-11-24 00:27:46 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s 172-237-154-236 coredns-668d6bf9bc-l5m2s eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali610973304d4 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Namespace="kube-system" Pod="coredns-668d6bf9bc-l5m2s" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-" Nov 24 00:28:16.874240 containerd[1559]: 2025-11-24 00:28:16.793 [INFO][4165] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Namespace="kube-system" Pod="coredns-668d6bf9bc-l5m2s" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" Nov 24 00:28:16.874240 containerd[1559]: 2025-11-24 00:28:16.820 [INFO][4176] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" HandleID="k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Workload="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.820 [INFO][4176] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" HandleID="k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Workload="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f220), Attrs:map[string]string{"namespace":"kube-system", "node":"172-237-154-236", "pod":"coredns-668d6bf9bc-l5m2s", "timestamp":"2025-11-24 00:28:16.820239115 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.821 [INFO][4176] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.821 [INFO][4176] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.821 [INFO][4176] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.832 [INFO][4176] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" host="172-237-154-236" Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.836 [INFO][4176] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.840 [INFO][4176] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.842 [INFO][4176] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.843 [INFO][4176] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:16.874401 containerd[1559]: 2025-11-24 00:28:16.843 [INFO][4176] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" host="172-237-154-236" Nov 24 00:28:16.874625 containerd[1559]: 2025-11-24 00:28:16.844 [INFO][4176] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1 Nov 24 00:28:16.874625 containerd[1559]: 2025-11-24 00:28:16.849 [INFO][4176] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" host="172-237-154-236" Nov 24 00:28:16.874625 containerd[1559]: 2025-11-24 00:28:16.853 [INFO][4176] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.130/26] block=192.168.40.128/26 handle="k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" host="172-237-154-236" Nov 24 00:28:16.874625 containerd[1559]: 2025-11-24 00:28:16.853 [INFO][4176] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.130/26] handle="k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" host="172-237-154-236" Nov 24 00:28:16.874625 containerd[1559]: 2025-11-24 00:28:16.853 [INFO][4176] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:16.874625 containerd[1559]: 2025-11-24 00:28:16.853 [INFO][4176] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.130/26] IPv6=[] ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" HandleID="k8s-pod-network.7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Workload="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" Nov 24 00:28:16.874741 containerd[1559]: 2025-11-24 00:28:16.856 [INFO][4165] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Namespace="kube-system" Pod="coredns-668d6bf9bc-l5m2s" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"2fada79a-43af-4af3-bfd2-ddde7f2c0d48", ResourceVersion:"823", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"coredns-668d6bf9bc-l5m2s", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.40.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali610973304d4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:16.874741 containerd[1559]: 2025-11-24 00:28:16.856 [INFO][4165] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.130/32] ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Namespace="kube-system" Pod="coredns-668d6bf9bc-l5m2s" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" Nov 24 00:28:16.874741 containerd[1559]: 2025-11-24 00:28:16.856 [INFO][4165] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali610973304d4 ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Namespace="kube-system" Pod="coredns-668d6bf9bc-l5m2s" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" Nov 24 00:28:16.874741 containerd[1559]: 2025-11-24 00:28:16.858 [INFO][4165] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Namespace="kube-system" Pod="coredns-668d6bf9bc-l5m2s" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" Nov 24 00:28:16.874741 containerd[1559]: 2025-11-24 00:28:16.858 [INFO][4165] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Namespace="kube-system" Pod="coredns-668d6bf9bc-l5m2s" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"2fada79a-43af-4af3-bfd2-ddde7f2c0d48", ResourceVersion:"823", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1", Pod:"coredns-668d6bf9bc-l5m2s", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.40.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali610973304d4", MAC:"42:da:43:1f:9b:76", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:16.874741 containerd[1559]: 2025-11-24 00:28:16.867 [INFO][4165] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" Namespace="kube-system" Pod="coredns-668d6bf9bc-l5m2s" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--l5m2s-eth0" Nov 24 00:28:16.900821 containerd[1559]: time="2025-11-24T00:28:16.900774415Z" level=info msg="connecting to shim 7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1" address="unix:///run/containerd/s/c53b532b027218c3b2262d838bc495622961d1fe793b30a8b84ef3c44f96a527" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:16.934042 systemd[1]: Started cri-containerd-7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1.scope - libcontainer container 7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1. Nov 24 00:28:16.988983 containerd[1559]: time="2025-11-24T00:28:16.988948049Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-l5m2s,Uid:2fada79a-43af-4af3-bfd2-ddde7f2c0d48,Namespace:kube-system,Attempt:0,} returns sandbox id \"7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1\"" Nov 24 00:28:16.992932 kubelet[2709]: E1124 00:28:16.992096 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:16.996860 containerd[1559]: time="2025-11-24T00:28:16.996763953Z" level=info msg="CreateContainer within sandbox \"7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 24 00:28:17.006595 containerd[1559]: time="2025-11-24T00:28:17.006574818Z" level=info msg="Container 5fbd4d253f306f5658a34d76400a105be692d045e47bdd11b938cdcb06eafb80: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:28:17.015122 containerd[1559]: time="2025-11-24T00:28:17.015057352Z" level=info msg="CreateContainer within sandbox \"7940bbc62ffea7306de441037e5282f4c94ca9ce5154d964a03c84f272f2b2c1\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"5fbd4d253f306f5658a34d76400a105be692d045e47bdd11b938cdcb06eafb80\"" Nov 24 00:28:17.017006 containerd[1559]: time="2025-11-24T00:28:17.015635573Z" level=info msg="StartContainer for \"5fbd4d253f306f5658a34d76400a105be692d045e47bdd11b938cdcb06eafb80\"" Nov 24 00:28:17.017006 containerd[1559]: time="2025-11-24T00:28:17.016292233Z" level=info msg="connecting to shim 5fbd4d253f306f5658a34d76400a105be692d045e47bdd11b938cdcb06eafb80" address="unix:///run/containerd/s/c53b532b027218c3b2262d838bc495622961d1fe793b30a8b84ef3c44f96a527" protocol=ttrpc version=3 Nov 24 00:28:17.036048 systemd[1]: Started cri-containerd-5fbd4d253f306f5658a34d76400a105be692d045e47bdd11b938cdcb06eafb80.scope - libcontainer container 5fbd4d253f306f5658a34d76400a105be692d045e47bdd11b938cdcb06eafb80. Nov 24 00:28:17.067936 containerd[1559]: time="2025-11-24T00:28:17.067887469Z" level=info msg="StartContainer for \"5fbd4d253f306f5658a34d76400a105be692d045e47bdd11b938cdcb06eafb80\" returns successfully" Nov 24 00:28:17.756492 containerd[1559]: time="2025-11-24T00:28:17.756409043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5b8f657c8-qdb6r,Uid:7208ef7a-cd09-4ec2-a692-f04c8a143f91,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:17.868504 systemd-networkd[1439]: cali14e81f7b5b9: Link UP Nov 24 00:28:17.870126 systemd-networkd[1439]: cali14e81f7b5b9: Gained carrier Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.795 [INFO][4282] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0 calico-kube-controllers-5b8f657c8- calico-system 7208ef7a-cd09-4ec2-a692-f04c8a143f91 829 0 2025-11-24 00:28:00 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5b8f657c8 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s 172-237-154-236 calico-kube-controllers-5b8f657c8-qdb6r eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali14e81f7b5b9 [] [] }} ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Namespace="calico-system" Pod="calico-kube-controllers-5b8f657c8-qdb6r" WorkloadEndpoint="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.795 [INFO][4282] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Namespace="calico-system" Pod="calico-kube-controllers-5b8f657c8-qdb6r" WorkloadEndpoint="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.831 [INFO][4294] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" HandleID="k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Workload="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.832 [INFO][4294] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" HandleID="k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Workload="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d4fe0), Attrs:map[string]string{"namespace":"calico-system", "node":"172-237-154-236", "pod":"calico-kube-controllers-5b8f657c8-qdb6r", "timestamp":"2025-11-24 00:28:17.831979621 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.832 [INFO][4294] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.832 [INFO][4294] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.832 [INFO][4294] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.840 [INFO][4294] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.844 [INFO][4294] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.847 [INFO][4294] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.849 [INFO][4294] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.850 [INFO][4294] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.850 [INFO][4294] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.851 [INFO][4294] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5 Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.855 [INFO][4294] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.860 [INFO][4294] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.131/26] block=192.168.40.128/26 handle="k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.861 [INFO][4294] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.131/26] handle="k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" host="172-237-154-236" Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.861 [INFO][4294] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:17.885673 containerd[1559]: 2025-11-24 00:28:17.861 [INFO][4294] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.131/26] IPv6=[] ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" HandleID="k8s-pod-network.b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Workload="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" Nov 24 00:28:17.886456 containerd[1559]: 2025-11-24 00:28:17.865 [INFO][4282] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Namespace="calico-system" Pod="calico-kube-controllers-5b8f657c8-qdb6r" WorkloadEndpoint="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0", GenerateName:"calico-kube-controllers-5b8f657c8-", Namespace:"calico-system", SelfLink:"", UID:"7208ef7a-cd09-4ec2-a692-f04c8a143f91", ResourceVersion:"829", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 28, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5b8f657c8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"calico-kube-controllers-5b8f657c8-qdb6r", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.40.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali14e81f7b5b9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:17.886456 containerd[1559]: 2025-11-24 00:28:17.865 [INFO][4282] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.131/32] ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Namespace="calico-system" Pod="calico-kube-controllers-5b8f657c8-qdb6r" WorkloadEndpoint="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" Nov 24 00:28:17.886456 containerd[1559]: 2025-11-24 00:28:17.865 [INFO][4282] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali14e81f7b5b9 ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Namespace="calico-system" Pod="calico-kube-controllers-5b8f657c8-qdb6r" WorkloadEndpoint="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" Nov 24 00:28:17.886456 containerd[1559]: 2025-11-24 00:28:17.869 [INFO][4282] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Namespace="calico-system" Pod="calico-kube-controllers-5b8f657c8-qdb6r" WorkloadEndpoint="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" Nov 24 00:28:17.886456 containerd[1559]: 2025-11-24 00:28:17.869 [INFO][4282] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Namespace="calico-system" Pod="calico-kube-controllers-5b8f657c8-qdb6r" WorkloadEndpoint="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0", GenerateName:"calico-kube-controllers-5b8f657c8-", Namespace:"calico-system", SelfLink:"", UID:"7208ef7a-cd09-4ec2-a692-f04c8a143f91", ResourceVersion:"829", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 28, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5b8f657c8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5", Pod:"calico-kube-controllers-5b8f657c8-qdb6r", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.40.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali14e81f7b5b9", MAC:"16:4b:59:f8:0b:08", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:17.886456 containerd[1559]: 2025-11-24 00:28:17.878 [INFO][4282] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" Namespace="calico-system" Pod="calico-kube-controllers-5b8f657c8-qdb6r" WorkloadEndpoint="172--237--154--236-k8s-calico--kube--controllers--5b8f657c8--qdb6r-eth0" Nov 24 00:28:17.916079 containerd[1559]: time="2025-11-24T00:28:17.916011443Z" level=info msg="connecting to shim b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5" address="unix:///run/containerd/s/4f65eef5fcada0788008db96d0e06a0ebb01ed880b13d0f5f8242a07eab0a0b5" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:17.922283 kubelet[2709]: E1124 00:28:17.922266 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:17.959253 systemd[1]: Started cri-containerd-b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5.scope - libcontainer container b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5. Nov 24 00:28:17.970007 kubelet[2709]: I1124 00:28:17.969606 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-l5m2s" podStartSLOduration=31.969589779 podStartE2EDuration="31.969589779s" podCreationTimestamp="2025-11-24 00:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:28:17.947043268 +0000 UTC m=+37.282851870" watchObservedRunningTime="2025-11-24 00:28:17.969589779 +0000 UTC m=+37.305398391" Nov 24 00:28:18.030985 containerd[1559]: time="2025-11-24T00:28:18.030334080Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5b8f657c8-qdb6r,Uid:7208ef7a-cd09-4ec2-a692-f04c8a143f91,Namespace:calico-system,Attempt:0,} returns sandbox id \"b6e68a2388ff3859e7442bc90a95a35ea5d3626614ebe4c4e0e935e701c688c5\"" Nov 24 00:28:18.033207 containerd[1559]: time="2025-11-24T00:28:18.033184891Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 24 00:28:18.167680 containerd[1559]: time="2025-11-24T00:28:18.167566148Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:18.168592 containerd[1559]: time="2025-11-24T00:28:18.168529549Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 24 00:28:18.168642 containerd[1559]: time="2025-11-24T00:28:18.168588049Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 24 00:28:18.169039 kubelet[2709]: E1124 00:28:18.168988 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:28:18.169210 kubelet[2709]: E1124 00:28:18.169141 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:28:18.169368 kubelet[2709]: E1124 00:28:18.169331 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5dwz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5b8f657c8-qdb6r_calico-system(7208ef7a-cd09-4ec2-a692-f04c8a143f91): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:18.171198 kubelet[2709]: E1124 00:28:18.171137 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:28:18.484283 systemd-networkd[1439]: cali610973304d4: Gained IPv6LL Nov 24 00:28:18.757135 containerd[1559]: time="2025-11-24T00:28:18.756939303Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7595d974c5-lnnqz,Uid:96698920-d098-4174-b103-f535ce1caea1,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:28:18.878110 systemd-networkd[1439]: cali48a93ffd10d: Link UP Nov 24 00:28:18.881033 systemd-networkd[1439]: cali48a93ffd10d: Gained carrier Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.799 [INFO][4362] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0 calico-apiserver-7595d974c5- calico-apiserver 96698920-d098-4174-b103-f535ce1caea1 826 0 2025-11-24 00:27:56 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7595d974c5 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s 172-237-154-236 calico-apiserver-7595d974c5-lnnqz eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali48a93ffd10d [] [] }} ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Namespace="calico-apiserver" Pod="calico-apiserver-7595d974c5-lnnqz" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.799 [INFO][4362] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Namespace="calico-apiserver" Pod="calico-apiserver-7595d974c5-lnnqz" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.828 [INFO][4375] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" HandleID="k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Workload="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.828 [INFO][4375] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" HandleID="k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Workload="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d4fe0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"172-237-154-236", "pod":"calico-apiserver-7595d974c5-lnnqz", "timestamp":"2025-11-24 00:28:18.828264178 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.828 [INFO][4375] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.828 [INFO][4375] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.828 [INFO][4375] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.834 [INFO][4375] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.839 [INFO][4375] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.845 [INFO][4375] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.848 [INFO][4375] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.851 [INFO][4375] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.851 [INFO][4375] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.852 [INFO][4375] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.858 [INFO][4375] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.864 [INFO][4375] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.132/26] block=192.168.40.128/26 handle="k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.864 [INFO][4375] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.132/26] handle="k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" host="172-237-154-236" Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.864 [INFO][4375] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:18.900569 containerd[1559]: 2025-11-24 00:28:18.864 [INFO][4375] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.132/26] IPv6=[] ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" HandleID="k8s-pod-network.6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Workload="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" Nov 24 00:28:18.902161 containerd[1559]: 2025-11-24 00:28:18.869 [INFO][4362] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Namespace="calico-apiserver" Pod="calico-apiserver-7595d974c5-lnnqz" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0", GenerateName:"calico-apiserver-7595d974c5-", Namespace:"calico-apiserver", SelfLink:"", UID:"96698920-d098-4174-b103-f535ce1caea1", ResourceVersion:"826", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7595d974c5", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"calico-apiserver-7595d974c5-lnnqz", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali48a93ffd10d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:18.902161 containerd[1559]: 2025-11-24 00:28:18.869 [INFO][4362] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.132/32] ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Namespace="calico-apiserver" Pod="calico-apiserver-7595d974c5-lnnqz" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" Nov 24 00:28:18.902161 containerd[1559]: 2025-11-24 00:28:18.869 [INFO][4362] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali48a93ffd10d ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Namespace="calico-apiserver" Pod="calico-apiserver-7595d974c5-lnnqz" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" Nov 24 00:28:18.902161 containerd[1559]: 2025-11-24 00:28:18.873 [INFO][4362] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Namespace="calico-apiserver" Pod="calico-apiserver-7595d974c5-lnnqz" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" Nov 24 00:28:18.902161 containerd[1559]: 2025-11-24 00:28:18.873 [INFO][4362] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Namespace="calico-apiserver" Pod="calico-apiserver-7595d974c5-lnnqz" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0", GenerateName:"calico-apiserver-7595d974c5-", Namespace:"calico-apiserver", SelfLink:"", UID:"96698920-d098-4174-b103-f535ce1caea1", ResourceVersion:"826", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7595d974c5", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa", Pod:"calico-apiserver-7595d974c5-lnnqz", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali48a93ffd10d", MAC:"aa:a3:67:c0:f9:49", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:18.902161 containerd[1559]: 2025-11-24 00:28:18.886 [INFO][4362] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" Namespace="calico-apiserver" Pod="calico-apiserver-7595d974c5-lnnqz" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--7595d974c5--lnnqz-eth0" Nov 24 00:28:18.925677 kubelet[2709]: E1124 00:28:18.925604 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:18.931302 kubelet[2709]: E1124 00:28:18.931267 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:28:18.932210 containerd[1559]: time="2025-11-24T00:28:18.931712150Z" level=info msg="connecting to shim 6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa" address="unix:///run/containerd/s/e27de2562a14f7912708c64ed280c2fb45f02569eac6c7db391f9e08939989bf" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:18.968043 systemd[1]: Started cri-containerd-6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa.scope - libcontainer container 6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa. Nov 24 00:28:19.027013 containerd[1559]: time="2025-11-24T00:28:19.026252167Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7595d974c5-lnnqz,Uid:96698920-d098-4174-b103-f535ce1caea1,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"6defcc375883b6f0458d73d65fac7ea59c510ccb51ba335d98f22b60003368aa\"" Nov 24 00:28:19.029068 containerd[1559]: time="2025-11-24T00:28:19.029005069Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:28:19.154110 containerd[1559]: time="2025-11-24T00:28:19.153896601Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:19.155313 containerd[1559]: time="2025-11-24T00:28:19.155232292Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:28:19.155313 containerd[1559]: time="2025-11-24T00:28:19.155301792Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:19.158933 kubelet[2709]: E1124 00:28:19.158874 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:19.158980 kubelet[2709]: E1124 00:28:19.158932 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:19.159104 kubelet[2709]: E1124 00:28:19.159043 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-69sxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7595d974c5-lnnqz_calico-apiserver(96698920-d098-4174-b103-f535ce1caea1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:19.160323 kubelet[2709]: E1124 00:28:19.160282 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:28:19.251117 systemd-networkd[1439]: cali14e81f7b5b9: Gained IPv6LL Nov 24 00:28:19.756357 kubelet[2709]: E1124 00:28:19.755982 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:19.757071 containerd[1559]: time="2025-11-24T00:28:19.757023322Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hj4d7,Uid:58c2ada9-58c2-46b7-8f97-99d17c62d364,Namespace:kube-system,Attempt:0,}" Nov 24 00:28:19.862341 systemd-networkd[1439]: cali5bc4a57f566: Link UP Nov 24 00:28:19.864062 systemd-networkd[1439]: cali5bc4a57f566: Gained carrier Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.796 [INFO][4434] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0 coredns-668d6bf9bc- kube-system 58c2ada9-58c2-46b7-8f97-99d17c62d364 817 0 2025-11-24 00:27:46 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s 172-237-154-236 coredns-668d6bf9bc-hj4d7 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali5bc4a57f566 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Namespace="kube-system" Pod="coredns-668d6bf9bc-hj4d7" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.796 [INFO][4434] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Namespace="kube-system" Pod="coredns-668d6bf9bc-hj4d7" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.819 [INFO][4445] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" HandleID="k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Workload="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.820 [INFO][4445] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" HandleID="k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Workload="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f230), Attrs:map[string]string{"namespace":"kube-system", "node":"172-237-154-236", "pod":"coredns-668d6bf9bc-hj4d7", "timestamp":"2025-11-24 00:28:19.819944344 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.820 [INFO][4445] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.820 [INFO][4445] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.820 [INFO][4445] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.826 [INFO][4445] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.834 [INFO][4445] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.837 [INFO][4445] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.839 [INFO][4445] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.841 [INFO][4445] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.841 [INFO][4445] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.842 [INFO][4445] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.847 [INFO][4445] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.852 [INFO][4445] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.133/26] block=192.168.40.128/26 handle="k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.852 [INFO][4445] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.133/26] handle="k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" host="172-237-154-236" Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.853 [INFO][4445] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:19.880639 containerd[1559]: 2025-11-24 00:28:19.853 [INFO][4445] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.133/26] IPv6=[] ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" HandleID="k8s-pod-network.364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Workload="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" Nov 24 00:28:19.881673 containerd[1559]: 2025-11-24 00:28:19.856 [INFO][4434] cni-plugin/k8s.go 418: Populated endpoint ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Namespace="kube-system" Pod="coredns-668d6bf9bc-hj4d7" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"58c2ada9-58c2-46b7-8f97-99d17c62d364", ResourceVersion:"817", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"coredns-668d6bf9bc-hj4d7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.40.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali5bc4a57f566", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:19.881673 containerd[1559]: 2025-11-24 00:28:19.856 [INFO][4434] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.133/32] ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Namespace="kube-system" Pod="coredns-668d6bf9bc-hj4d7" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" Nov 24 00:28:19.881673 containerd[1559]: 2025-11-24 00:28:19.856 [INFO][4434] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5bc4a57f566 ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Namespace="kube-system" Pod="coredns-668d6bf9bc-hj4d7" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" Nov 24 00:28:19.881673 containerd[1559]: 2025-11-24 00:28:19.867 [INFO][4434] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Namespace="kube-system" Pod="coredns-668d6bf9bc-hj4d7" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" Nov 24 00:28:19.881673 containerd[1559]: 2025-11-24 00:28:19.868 [INFO][4434] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Namespace="kube-system" Pod="coredns-668d6bf9bc-hj4d7" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"58c2ada9-58c2-46b7-8f97-99d17c62d364", ResourceVersion:"817", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad", Pod:"coredns-668d6bf9bc-hj4d7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.40.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali5bc4a57f566", MAC:"5a:2f:99:d0:06:e5", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:19.881673 containerd[1559]: 2025-11-24 00:28:19.876 [INFO][4434] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" Namespace="kube-system" Pod="coredns-668d6bf9bc-hj4d7" WorkloadEndpoint="172--237--154--236-k8s-coredns--668d6bf9bc--hj4d7-eth0" Nov 24 00:28:19.905011 containerd[1559]: time="2025-11-24T00:28:19.904981136Z" level=info msg="connecting to shim 364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad" address="unix:///run/containerd/s/a4a0dff3bcfdde687e40aee33c53b63aa9588ce0de2585bfd6727ec81a27899b" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:19.935708 kubelet[2709]: E1124 00:28:19.935146 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:19.936139 kubelet[2709]: E1124 00:28:19.935290 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:28:19.936227 kubelet[2709]: E1124 00:28:19.935526 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:28:19.938104 systemd[1]: Started cri-containerd-364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad.scope - libcontainer container 364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad. Nov 24 00:28:20.016130 containerd[1559]: time="2025-11-24T00:28:20.015792782Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hj4d7,Uid:58c2ada9-58c2-46b7-8f97-99d17c62d364,Namespace:kube-system,Attempt:0,} returns sandbox id \"364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad\"" Nov 24 00:28:20.017455 kubelet[2709]: E1124 00:28:20.017137 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:20.020328 containerd[1559]: time="2025-11-24T00:28:20.020303394Z" level=info msg="CreateContainer within sandbox \"364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 24 00:28:20.029358 containerd[1559]: time="2025-11-24T00:28:20.028437458Z" level=info msg="Container 4e4f7938619a1a63136601c7772478df3c29162274c79834c328cfe27145fd0c: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:28:20.039525 containerd[1559]: time="2025-11-24T00:28:20.039503404Z" level=info msg="CreateContainer within sandbox \"364120b35dfa9f3e04a15cdac34a0bbe947a13b90c743da87ebde217ce2d4aad\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4e4f7938619a1a63136601c7772478df3c29162274c79834c328cfe27145fd0c\"" Nov 24 00:28:20.040226 containerd[1559]: time="2025-11-24T00:28:20.040210384Z" level=info msg="StartContainer for \"4e4f7938619a1a63136601c7772478df3c29162274c79834c328cfe27145fd0c\"" Nov 24 00:28:20.041186 containerd[1559]: time="2025-11-24T00:28:20.041127314Z" level=info msg="connecting to shim 4e4f7938619a1a63136601c7772478df3c29162274c79834c328cfe27145fd0c" address="unix:///run/containerd/s/a4a0dff3bcfdde687e40aee33c53b63aa9588ce0de2585bfd6727ec81a27899b" protocol=ttrpc version=3 Nov 24 00:28:20.068054 systemd[1]: Started cri-containerd-4e4f7938619a1a63136601c7772478df3c29162274c79834c328cfe27145fd0c.scope - libcontainer container 4e4f7938619a1a63136601c7772478df3c29162274c79834c328cfe27145fd0c. Nov 24 00:28:20.106213 containerd[1559]: time="2025-11-24T00:28:20.106136917Z" level=info msg="StartContainer for \"4e4f7938619a1a63136601c7772478df3c29162274c79834c328cfe27145fd0c\" returns successfully" Nov 24 00:28:20.275269 systemd-networkd[1439]: cali48a93ffd10d: Gained IPv6LL Nov 24 00:28:20.757109 containerd[1559]: time="2025-11-24T00:28:20.756858612Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-g9zxr,Uid:be91967a-4242-4c33-899c-c3fa81f6e4dc,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:20.757109 containerd[1559]: time="2025-11-24T00:28:20.756936312Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-844c564bfd-tnmdj,Uid:c6030a3d-68f0-4144-8d41-801a65d0dbd4,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:28:20.757645 containerd[1559]: time="2025-11-24T00:28:20.756860542Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5vjmh,Uid:d80649f2-a0ab-46f8-9106-ed888d934bff,Namespace:calico-system,Attempt:0,}" Nov 24 00:28:20.922339 systemd-networkd[1439]: calid709f8d2f86: Link UP Nov 24 00:28:20.925991 systemd-networkd[1439]: calid709f8d2f86: Gained carrier Nov 24 00:28:20.940357 kubelet[2709]: E1124 00:28:20.940330 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:20.944080 kubelet[2709]: E1124 00:28:20.943967 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.827 [INFO][4559] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0 calico-apiserver-844c564bfd- calico-apiserver c6030a3d-68f0-4144-8d41-801a65d0dbd4 828 0 2025-11-24 00:27:55 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:844c564bfd projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s 172-237-154-236 calico-apiserver-844c564bfd-tnmdj eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calid709f8d2f86 [] [] }} ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-tnmdj" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.828 [INFO][4559] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-tnmdj" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.879 [INFO][4579] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" HandleID="k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Workload="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.880 [INFO][4579] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" HandleID="k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Workload="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003aec60), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"172-237-154-236", "pod":"calico-apiserver-844c564bfd-tnmdj", "timestamp":"2025-11-24 00:28:20.879585063 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.881 [INFO][4579] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.881 [INFO][4579] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.881 [INFO][4579] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.891 [INFO][4579] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.896 [INFO][4579] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.900 [INFO][4579] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.902 [INFO][4579] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.904 [INFO][4579] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.904 [INFO][4579] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.905 [INFO][4579] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4 Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.908 [INFO][4579] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.913 [INFO][4579] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.134/26] block=192.168.40.128/26 handle="k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.913 [INFO][4579] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.134/26] handle="k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" host="172-237-154-236" Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.913 [INFO][4579] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:20.946477 containerd[1559]: 2025-11-24 00:28:20.913 [INFO][4579] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.134/26] IPv6=[] ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" HandleID="k8s-pod-network.8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Workload="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" Nov 24 00:28:20.947677 containerd[1559]: 2025-11-24 00:28:20.916 [INFO][4559] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-tnmdj" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0", GenerateName:"calico-apiserver-844c564bfd-", Namespace:"calico-apiserver", SelfLink:"", UID:"c6030a3d-68f0-4144-8d41-801a65d0dbd4", ResourceVersion:"828", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"844c564bfd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"calico-apiserver-844c564bfd-tnmdj", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calid709f8d2f86", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:20.947677 containerd[1559]: 2025-11-24 00:28:20.916 [INFO][4559] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.134/32] ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-tnmdj" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" Nov 24 00:28:20.947677 containerd[1559]: 2025-11-24 00:28:20.916 [INFO][4559] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid709f8d2f86 ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-tnmdj" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" Nov 24 00:28:20.947677 containerd[1559]: 2025-11-24 00:28:20.926 [INFO][4559] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-tnmdj" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" Nov 24 00:28:20.947677 containerd[1559]: 2025-11-24 00:28:20.926 [INFO][4559] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-tnmdj" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0", GenerateName:"calico-apiserver-844c564bfd-", Namespace:"calico-apiserver", SelfLink:"", UID:"c6030a3d-68f0-4144-8d41-801a65d0dbd4", ResourceVersion:"828", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"844c564bfd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4", Pod:"calico-apiserver-844c564bfd-tnmdj", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calid709f8d2f86", MAC:"ba:a6:97:ce:c7:40", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:20.947677 containerd[1559]: 2025-11-24 00:28:20.940 [INFO][4559] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-tnmdj" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--tnmdj-eth0" Nov 24 00:28:20.956483 kubelet[2709]: I1124 00:28:20.956444 2709 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-hj4d7" podStartSLOduration=34.956431512 podStartE2EDuration="34.956431512s" podCreationTimestamp="2025-11-24 00:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:28:20.955725411 +0000 UTC m=+40.291534033" watchObservedRunningTime="2025-11-24 00:28:20.956431512 +0000 UTC m=+40.292240134" Nov 24 00:28:21.002216 containerd[1559]: time="2025-11-24T00:28:21.001835244Z" level=info msg="connecting to shim 8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4" address="unix:///run/containerd/s/a3511f0da37acefc458f39ff89ec3d91bd85f2b7a06d55080977b1c62f63d87d" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:21.051082 systemd[1]: Started cri-containerd-8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4.scope - libcontainer container 8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4. Nov 24 00:28:21.077549 systemd-networkd[1439]: cali824827a9154: Link UP Nov 24 00:28:21.078894 systemd-networkd[1439]: cali824827a9154: Gained carrier Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:20.828 [INFO][4549] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-csi--node--driver--5vjmh-eth0 csi-node-driver- calico-system d80649f2-a0ab-46f8-9106-ed888d934bff 722 0 2025-11-24 00:28:00 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s 172-237-154-236 csi-node-driver-5vjmh eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali824827a9154 [] [] }} ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Namespace="calico-system" Pod="csi-node-driver-5vjmh" WorkloadEndpoint="172--237--154--236-k8s-csi--node--driver--5vjmh-" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:20.828 [INFO][4549] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Namespace="calico-system" Pod="csi-node-driver-5vjmh" WorkloadEndpoint="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:20.883 [INFO][4581] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" HandleID="k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Workload="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:20.883 [INFO][4581] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" HandleID="k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Workload="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000399ae0), Attrs:map[string]string{"namespace":"calico-system", "node":"172-237-154-236", "pod":"csi-node-driver-5vjmh", "timestamp":"2025-11-24 00:28:20.883852345 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:20.884 [INFO][4581] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:20.913 [INFO][4581] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:20.913 [INFO][4581] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:20.995 [INFO][4581] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.019 [INFO][4581] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.031 [INFO][4581] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.037 [INFO][4581] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.039 [INFO][4581] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.039 [INFO][4581] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.043 [INFO][4581] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.050 [INFO][4581] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.060 [INFO][4581] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.135/26] block=192.168.40.128/26 handle="k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.060 [INFO][4581] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.135/26] handle="k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" host="172-237-154-236" Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.060 [INFO][4581] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:21.100084 containerd[1559]: 2025-11-24 00:28:21.060 [INFO][4581] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.135/26] IPv6=[] ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" HandleID="k8s-pod-network.7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Workload="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" Nov 24 00:28:21.100674 containerd[1559]: 2025-11-24 00:28:21.068 [INFO][4549] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Namespace="calico-system" Pod="csi-node-driver-5vjmh" WorkloadEndpoint="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-csi--node--driver--5vjmh-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d80649f2-a0ab-46f8-9106-ed888d934bff", ResourceVersion:"722", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 28, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"csi-node-driver-5vjmh", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.40.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali824827a9154", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:21.100674 containerd[1559]: 2025-11-24 00:28:21.068 [INFO][4549] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.135/32] ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Namespace="calico-system" Pod="csi-node-driver-5vjmh" WorkloadEndpoint="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" Nov 24 00:28:21.100674 containerd[1559]: 2025-11-24 00:28:21.069 [INFO][4549] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali824827a9154 ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Namespace="calico-system" Pod="csi-node-driver-5vjmh" WorkloadEndpoint="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" Nov 24 00:28:21.100674 containerd[1559]: 2025-11-24 00:28:21.077 [INFO][4549] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Namespace="calico-system" Pod="csi-node-driver-5vjmh" WorkloadEndpoint="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" Nov 24 00:28:21.100674 containerd[1559]: 2025-11-24 00:28:21.077 [INFO][4549] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Namespace="calico-system" Pod="csi-node-driver-5vjmh" WorkloadEndpoint="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-csi--node--driver--5vjmh-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d80649f2-a0ab-46f8-9106-ed888d934bff", ResourceVersion:"722", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 28, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a", Pod:"csi-node-driver-5vjmh", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.40.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali824827a9154", MAC:"ba:ee:fe:bd:41:76", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:21.100674 containerd[1559]: 2025-11-24 00:28:21.095 [INFO][4549] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" Namespace="calico-system" Pod="csi-node-driver-5vjmh" WorkloadEndpoint="172--237--154--236-k8s-csi--node--driver--5vjmh-eth0" Nov 24 00:28:21.150987 containerd[1559]: time="2025-11-24T00:28:21.149445288Z" level=info msg="connecting to shim 7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a" address="unix:///run/containerd/s/0fdb4eab51908f0b1f5b280d5cdd34bcdf18563d502a67644f0602b7d572ea02" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:21.186053 systemd-networkd[1439]: cali75b1ea2683f: Link UP Nov 24 00:28:21.187028 systemd-networkd[1439]: cali75b1ea2683f: Gained carrier Nov 24 00:28:21.217126 systemd[1]: Started cri-containerd-7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a.scope - libcontainer container 7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a. Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:20.842 [INFO][4541] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0 goldmane-666569f655- calico-system be91967a-4242-4c33-899c-c3fa81f6e4dc 825 0 2025-11-24 00:27:58 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s 172-237-154-236 goldmane-666569f655-g9zxr eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali75b1ea2683f [] [] }} ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Namespace="calico-system" Pod="goldmane-666569f655-g9zxr" WorkloadEndpoint="172--237--154--236-k8s-goldmane--666569f655--g9zxr-" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:20.843 [INFO][4541] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Namespace="calico-system" Pod="goldmane-666569f655-g9zxr" WorkloadEndpoint="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:20.887 [INFO][4589] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" HandleID="k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Workload="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:20.888 [INFO][4589] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" HandleID="k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Workload="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002cf020), Attrs:map[string]string{"namespace":"calico-system", "node":"172-237-154-236", "pod":"goldmane-666569f655-g9zxr", "timestamp":"2025-11-24 00:28:20.887891547 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:20.889 [INFO][4589] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.061 [INFO][4589] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.062 [INFO][4589] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.095 [INFO][4589] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.121 [INFO][4589] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.129 [INFO][4589] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.133 [INFO][4589] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.138 [INFO][4589] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.139 [INFO][4589] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.143 [INFO][4589] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16 Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.152 [INFO][4589] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.166 [INFO][4589] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.136/26] block=192.168.40.128/26 handle="k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.166 [INFO][4589] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.136/26] handle="k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" host="172-237-154-236" Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.166 [INFO][4589] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:21.220117 containerd[1559]: 2025-11-24 00:28:21.166 [INFO][4589] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.136/26] IPv6=[] ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" HandleID="k8s-pod-network.73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Workload="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" Nov 24 00:28:21.221275 containerd[1559]: 2025-11-24 00:28:21.177 [INFO][4541] cni-plugin/k8s.go 418: Populated endpoint ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Namespace="calico-system" Pod="goldmane-666569f655-g9zxr" WorkloadEndpoint="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"be91967a-4242-4c33-899c-c3fa81f6e4dc", ResourceVersion:"825", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"goldmane-666569f655-g9zxr", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.40.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali75b1ea2683f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:21.221275 containerd[1559]: 2025-11-24 00:28:21.178 [INFO][4541] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.136/32] ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Namespace="calico-system" Pod="goldmane-666569f655-g9zxr" WorkloadEndpoint="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" Nov 24 00:28:21.221275 containerd[1559]: 2025-11-24 00:28:21.179 [INFO][4541] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali75b1ea2683f ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Namespace="calico-system" Pod="goldmane-666569f655-g9zxr" WorkloadEndpoint="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" Nov 24 00:28:21.221275 containerd[1559]: 2025-11-24 00:28:21.192 [INFO][4541] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Namespace="calico-system" Pod="goldmane-666569f655-g9zxr" WorkloadEndpoint="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" Nov 24 00:28:21.221275 containerd[1559]: 2025-11-24 00:28:21.195 [INFO][4541] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Namespace="calico-system" Pod="goldmane-666569f655-g9zxr" WorkloadEndpoint="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"be91967a-4242-4c33-899c-c3fa81f6e4dc", ResourceVersion:"825", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16", Pod:"goldmane-666569f655-g9zxr", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.40.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali75b1ea2683f", MAC:"26:9f:d2:b9:3e:b1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:21.221275 containerd[1559]: 2025-11-24 00:28:21.210 [INFO][4541] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" Namespace="calico-system" Pod="goldmane-666569f655-g9zxr" WorkloadEndpoint="172--237--154--236-k8s-goldmane--666569f655--g9zxr-eth0" Nov 24 00:28:21.224372 containerd[1559]: time="2025-11-24T00:28:21.224335566Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-844c564bfd-tnmdj,Uid:c6030a3d-68f0-4144-8d41-801a65d0dbd4,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"8ca3786124bba1392d36a10f5c50853774d02de80347886b6437fb06487e60a4\"" Nov 24 00:28:21.226898 containerd[1559]: time="2025-11-24T00:28:21.226845317Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:28:21.248683 containerd[1559]: time="2025-11-24T00:28:21.248635758Z" level=info msg="connecting to shim 73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16" address="unix:///run/containerd/s/0dfc8a845f3754d6937ecf7b229909d528c53d5aafba50310dfe4dc0d3ee6208" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:21.295153 systemd[1]: Started cri-containerd-73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16.scope - libcontainer container 73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16. Nov 24 00:28:21.299120 containerd[1559]: time="2025-11-24T00:28:21.299078063Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5vjmh,Uid:d80649f2-a0ab-46f8-9106-ed888d934bff,Namespace:calico-system,Attempt:0,} returns sandbox id \"7bd8619407725bc4f2f0209f3a2ba2c9820a11c71ee9d683d4191a2c3974979a\"" Nov 24 00:28:21.350568 containerd[1559]: time="2025-11-24T00:28:21.350401509Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:21.351767 containerd[1559]: time="2025-11-24T00:28:21.351682229Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:28:21.351767 containerd[1559]: time="2025-11-24T00:28:21.351739909Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:21.352318 kubelet[2709]: E1124 00:28:21.352281 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:21.352418 kubelet[2709]: E1124 00:28:21.352386 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:21.352772 kubelet[2709]: E1124 00:28:21.352610 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h7tpq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-844c564bfd-tnmdj_calico-apiserver(c6030a3d-68f0-4144-8d41-801a65d0dbd4): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:21.353381 containerd[1559]: time="2025-11-24T00:28:21.353277410Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 24 00:28:21.354270 kubelet[2709]: E1124 00:28:21.354220 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:28:21.360841 containerd[1559]: time="2025-11-24T00:28:21.360819434Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-g9zxr,Uid:be91967a-4242-4c33-899c-c3fa81f6e4dc,Namespace:calico-system,Attempt:0,} returns sandbox id \"73e47a92f3f18feabd1d1da9cbb442ac011d280986ddcb6e44d8a5845920bd16\"" Nov 24 00:28:21.478855 containerd[1559]: time="2025-11-24T00:28:21.478769183Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:21.480210 containerd[1559]: time="2025-11-24T00:28:21.480160763Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 24 00:28:21.480345 containerd[1559]: time="2025-11-24T00:28:21.480170143Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 24 00:28:21.480459 kubelet[2709]: E1124 00:28:21.480423 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:28:21.480518 kubelet[2709]: E1124 00:28:21.480475 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:28:21.480724 kubelet[2709]: E1124 00:28:21.480676 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:21.481155 containerd[1559]: time="2025-11-24T00:28:21.481130274Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 24 00:28:21.613106 containerd[1559]: time="2025-11-24T00:28:21.612859360Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:21.614167 containerd[1559]: time="2025-11-24T00:28:21.614121460Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 24 00:28:21.614234 containerd[1559]: time="2025-11-24T00:28:21.614224420Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:21.614638 kubelet[2709]: E1124 00:28:21.614549 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:28:21.614638 kubelet[2709]: E1124 00:28:21.614632 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:28:21.615317 kubelet[2709]: E1124 00:28:21.615166 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m87kj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-g9zxr_calico-system(be91967a-4242-4c33-899c-c3fa81f6e4dc): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:21.616345 containerd[1559]: time="2025-11-24T00:28:21.616290571Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 24 00:28:21.616765 kubelet[2709]: E1124 00:28:21.616647 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:28:21.744277 containerd[1559]: time="2025-11-24T00:28:21.744211945Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:21.745348 containerd[1559]: time="2025-11-24T00:28:21.745277286Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 24 00:28:21.745348 containerd[1559]: time="2025-11-24T00:28:21.745313476Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 24 00:28:21.745639 kubelet[2709]: E1124 00:28:21.745574 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:28:21.745692 kubelet[2709]: E1124 00:28:21.745646 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:28:21.746078 kubelet[2709]: E1124 00:28:21.745995 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:21.747512 kubelet[2709]: E1124 00:28:21.747448 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:21.755660 containerd[1559]: time="2025-11-24T00:28:21.755638531Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-844c564bfd-cldcl,Uid:d5afe1af-07b4-4767-ad04-88dc81c2e5e8,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:28:21.859410 systemd-networkd[1439]: cali1fb101012e5: Link UP Nov 24 00:28:21.862251 systemd-networkd[1439]: cali1fb101012e5: Gained carrier Nov 24 00:28:21.875079 systemd-networkd[1439]: cali5bc4a57f566: Gained IPv6LL Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.796 [INFO][4774] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0 calico-apiserver-844c564bfd- calico-apiserver d5afe1af-07b4-4767-ad04-88dc81c2e5e8 827 0 2025-11-24 00:27:55 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:844c564bfd projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s 172-237-154-236 calico-apiserver-844c564bfd-cldcl eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali1fb101012e5 [] [] }} ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-cldcl" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.796 [INFO][4774] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-cldcl" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.822 [INFO][4785] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" HandleID="k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Workload="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.822 [INFO][4785] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" HandleID="k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Workload="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f210), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"172-237-154-236", "pod":"calico-apiserver-844c564bfd-cldcl", "timestamp":"2025-11-24 00:28:21.822201224 +0000 UTC"}, Hostname:"172-237-154-236", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.822 [INFO][4785] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.822 [INFO][4785] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.822 [INFO][4785] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-237-154-236' Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.828 [INFO][4785] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.833 [INFO][4785] ipam/ipam.go 394: Looking up existing affinities for host host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.837 [INFO][4785] ipam/ipam.go 511: Trying affinity for 192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.838 [INFO][4785] ipam/ipam.go 158: Attempting to load block cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.841 [INFO][4785] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.40.128/26 host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.841 [INFO][4785] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.40.128/26 handle="k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.842 [INFO][4785] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4 Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.846 [INFO][4785] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.40.128/26 handle="k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.853 [INFO][4785] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.40.137/26] block=192.168.40.128/26 handle="k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.853 [INFO][4785] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.40.137/26] handle="k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" host="172-237-154-236" Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.853 [INFO][4785] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:28:21.883462 containerd[1559]: 2025-11-24 00:28:21.853 [INFO][4785] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.40.137/26] IPv6=[] ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" HandleID="k8s-pod-network.5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Workload="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" Nov 24 00:28:21.884252 containerd[1559]: 2025-11-24 00:28:21.856 [INFO][4774] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-cldcl" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0", GenerateName:"calico-apiserver-844c564bfd-", Namespace:"calico-apiserver", SelfLink:"", UID:"d5afe1af-07b4-4767-ad04-88dc81c2e5e8", ResourceVersion:"827", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"844c564bfd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"", Pod:"calico-apiserver-844c564bfd-cldcl", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.137/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali1fb101012e5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:21.884252 containerd[1559]: 2025-11-24 00:28:21.856 [INFO][4774] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.40.137/32] ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-cldcl" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" Nov 24 00:28:21.884252 containerd[1559]: 2025-11-24 00:28:21.856 [INFO][4774] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1fb101012e5 ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-cldcl" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" Nov 24 00:28:21.884252 containerd[1559]: 2025-11-24 00:28:21.863 [INFO][4774] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-cldcl" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" Nov 24 00:28:21.884252 containerd[1559]: 2025-11-24 00:28:21.863 [INFO][4774] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-cldcl" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0", GenerateName:"calico-apiserver-844c564bfd-", Namespace:"calico-apiserver", SelfLink:"", UID:"d5afe1af-07b4-4767-ad04-88dc81c2e5e8", ResourceVersion:"827", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 27, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"844c564bfd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-237-154-236", ContainerID:"5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4", Pod:"calico-apiserver-844c564bfd-cldcl", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.137/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali1fb101012e5", MAC:"c2:c3:5a:e0:b7:3a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:28:21.884252 containerd[1559]: 2025-11-24 00:28:21.872 [INFO][4774] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" Namespace="calico-apiserver" Pod="calico-apiserver-844c564bfd-cldcl" WorkloadEndpoint="172--237--154--236-k8s-calico--apiserver--844c564bfd--cldcl-eth0" Nov 24 00:28:21.923156 containerd[1559]: time="2025-11-24T00:28:21.923067025Z" level=info msg="connecting to shim 5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4" address="unix:///run/containerd/s/feb242e44c77fe170a65311ab00591226ba6cf47979b73a229f7e13ccd262612" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:28:21.959299 kubelet[2709]: E1124 00:28:21.958800 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:28:21.959159 systemd[1]: Started cri-containerd-5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4.scope - libcontainer container 5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4. Nov 24 00:28:21.965644 kubelet[2709]: E1124 00:28:21.965529 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:21.966327 kubelet[2709]: E1124 00:28:21.966310 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:21.967071 kubelet[2709]: E1124 00:28:21.967053 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:28:22.053780 containerd[1559]: time="2025-11-24T00:28:22.053532010Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-844c564bfd-cldcl,Uid:d5afe1af-07b4-4767-ad04-88dc81c2e5e8,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"5e27cfa68db777140a4aa3a72a4fa24fd9e6e47ce83953cdaa03732b9bc83bb4\"" Nov 24 00:28:22.057629 containerd[1559]: time="2025-11-24T00:28:22.056138251Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:28:22.183204 containerd[1559]: time="2025-11-24T00:28:22.183081055Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:22.184864 containerd[1559]: time="2025-11-24T00:28:22.184835485Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:28:22.185059 containerd[1559]: time="2025-11-24T00:28:22.184952726Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:22.185231 kubelet[2709]: E1124 00:28:22.185178 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:22.185231 kubelet[2709]: E1124 00:28:22.185228 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:22.185445 kubelet[2709]: E1124 00:28:22.185397 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vrxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-844c564bfd-cldcl_calico-apiserver(d5afe1af-07b4-4767-ad04-88dc81c2e5e8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:22.186959 kubelet[2709]: E1124 00:28:22.186686 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:28:22.387479 systemd-networkd[1439]: cali824827a9154: Gained IPv6LL Nov 24 00:28:22.388424 systemd-networkd[1439]: calid709f8d2f86: Gained IPv6LL Nov 24 00:28:22.969634 kubelet[2709]: E1124 00:28:22.969566 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:28:22.970087 kubelet[2709]: E1124 00:28:22.969943 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:22.971202 kubelet[2709]: E1124 00:28:22.970465 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:28:22.971202 kubelet[2709]: E1124 00:28:22.970532 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:28:22.971202 kubelet[2709]: E1124 00:28:22.970572 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:23.155109 systemd-networkd[1439]: cali75b1ea2683f: Gained IPv6LL Nov 24 00:28:23.539139 systemd-networkd[1439]: cali1fb101012e5: Gained IPv6LL Nov 24 00:28:23.972530 kubelet[2709]: E1124 00:28:23.972463 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:28:26.763076 containerd[1559]: time="2025-11-24T00:28:26.760695070Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 24 00:28:26.892884 containerd[1559]: time="2025-11-24T00:28:26.892823233Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:26.893884 containerd[1559]: time="2025-11-24T00:28:26.893857226Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 24 00:28:26.893985 containerd[1559]: time="2025-11-24T00:28:26.893943796Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 24 00:28:26.894092 kubelet[2709]: E1124 00:28:26.894055 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:28:26.894414 kubelet[2709]: E1124 00:28:26.894102 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:28:26.894414 kubelet[2709]: E1124 00:28:26.894191 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:4781a789a9cf40d795f128549fb7a9fc,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2q9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-67b5f8fb49-hhhrv_calico-system(64ccd4c8-759d-46f3-b211-0873db45cfc8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:26.896682 containerd[1559]: time="2025-11-24T00:28:26.896653352Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 24 00:28:27.024203 containerd[1559]: time="2025-11-24T00:28:27.024061608Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:27.025383 containerd[1559]: time="2025-11-24T00:28:27.025325724Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 24 00:28:27.025518 containerd[1559]: time="2025-11-24T00:28:27.025371959Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 24 00:28:27.029277 kubelet[2709]: E1124 00:28:27.025610 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:28:27.029469 kubelet[2709]: E1124 00:28:27.029387 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:28:27.029628 kubelet[2709]: E1124 00:28:27.029578 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m2q9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-67b5f8fb49-hhhrv_calico-system(64ccd4c8-759d-46f3-b211-0873db45cfc8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:27.031155 kubelet[2709]: E1124 00:28:27.031079 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:28:33.757535 containerd[1559]: time="2025-11-24T00:28:33.757467301Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 24 00:28:33.895671 containerd[1559]: time="2025-11-24T00:28:33.895585778Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:33.896951 containerd[1559]: time="2025-11-24T00:28:33.896776403Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 24 00:28:33.896951 containerd[1559]: time="2025-11-24T00:28:33.896836039Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:33.897286 kubelet[2709]: E1124 00:28:33.897227 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:28:33.897753 kubelet[2709]: E1124 00:28:33.897295 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:28:33.897753 kubelet[2709]: E1124 00:28:33.897595 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m87kj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-g9zxr_calico-system(be91967a-4242-4c33-899c-c3fa81f6e4dc): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:33.898751 containerd[1559]: time="2025-11-24T00:28:33.898386768Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:28:33.899010 kubelet[2709]: E1124 00:28:33.898685 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:28:34.038145 containerd[1559]: time="2025-11-24T00:28:34.037999979Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:34.038818 containerd[1559]: time="2025-11-24T00:28:34.038788037Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:28:34.038899 containerd[1559]: time="2025-11-24T00:28:34.038840723Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:34.038985 kubelet[2709]: E1124 00:28:34.038946 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:34.038985 kubelet[2709]: E1124 00:28:34.038975 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:34.039409 kubelet[2709]: E1124 00:28:34.039362 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-69sxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7595d974c5-lnnqz_calico-apiserver(96698920-d098-4174-b103-f535ce1caea1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:34.041020 kubelet[2709]: E1124 00:28:34.040582 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:28:34.756790 containerd[1559]: time="2025-11-24T00:28:34.756590744Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 24 00:28:34.888333 containerd[1559]: time="2025-11-24T00:28:34.888267063Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:34.889208 containerd[1559]: time="2025-11-24T00:28:34.889154674Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 24 00:28:34.889263 containerd[1559]: time="2025-11-24T00:28:34.889229979Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 24 00:28:34.889486 kubelet[2709]: E1124 00:28:34.889415 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:28:34.889486 kubelet[2709]: E1124 00:28:34.889477 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:28:34.889769 kubelet[2709]: E1124 00:28:34.889723 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5dwz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5b8f657c8-qdb6r_calico-system(7208ef7a-cd09-4ec2-a692-f04c8a143f91): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:34.890229 containerd[1559]: time="2025-11-24T00:28:34.890156616Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:28:34.890874 kubelet[2709]: E1124 00:28:34.890844 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:28:35.013748 containerd[1559]: time="2025-11-24T00:28:35.013591703Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:35.015119 containerd[1559]: time="2025-11-24T00:28:35.014954098Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:35.015194 containerd[1559]: time="2025-11-24T00:28:35.014992815Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:28:35.015361 kubelet[2709]: E1124 00:28:35.015326 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:35.015819 kubelet[2709]: E1124 00:28:35.015368 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:35.015819 kubelet[2709]: E1124 00:28:35.015484 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h7tpq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-844c564bfd-tnmdj_calico-apiserver(c6030a3d-68f0-4144-8d41-801a65d0dbd4): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:35.016682 kubelet[2709]: E1124 00:28:35.016655 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:28:36.757739 containerd[1559]: time="2025-11-24T00:28:36.757494614Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 24 00:28:36.890590 containerd[1559]: time="2025-11-24T00:28:36.890507472Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:36.891562 containerd[1559]: time="2025-11-24T00:28:36.891505003Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 24 00:28:36.891562 containerd[1559]: time="2025-11-24T00:28:36.891586648Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 24 00:28:36.892117 kubelet[2709]: E1124 00:28:36.891745 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:28:36.892117 kubelet[2709]: E1124 00:28:36.891801 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:28:36.892117 kubelet[2709]: E1124 00:28:36.892054 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:36.893212 containerd[1559]: time="2025-11-24T00:28:36.892858363Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:28:37.021965 containerd[1559]: time="2025-11-24T00:28:37.021795131Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:37.023157 containerd[1559]: time="2025-11-24T00:28:37.023080320Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:28:37.023157 containerd[1559]: time="2025-11-24T00:28:37.023158786Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:37.023371 kubelet[2709]: E1124 00:28:37.023283 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:37.023371 kubelet[2709]: E1124 00:28:37.023342 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:37.023801 kubelet[2709]: E1124 00:28:37.023755 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vrxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-844c564bfd-cldcl_calico-apiserver(d5afe1af-07b4-4767-ad04-88dc81c2e5e8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:37.024345 containerd[1559]: time="2025-11-24T00:28:37.024274855Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 24 00:28:37.025922 kubelet[2709]: E1124 00:28:37.025863 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:28:37.162595 containerd[1559]: time="2025-11-24T00:28:37.162527087Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:37.163866 containerd[1559]: time="2025-11-24T00:28:37.163825235Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 24 00:28:37.164010 containerd[1559]: time="2025-11-24T00:28:37.163946578Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 24 00:28:37.164220 kubelet[2709]: E1124 00:28:37.164172 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:28:37.164305 kubelet[2709]: E1124 00:28:37.164245 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:28:37.164452 kubelet[2709]: E1124 00:28:37.164404 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:37.166341 kubelet[2709]: E1124 00:28:37.166290 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:39.756775 kubelet[2709]: E1124 00:28:39.756645 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:28:41.991013 kubelet[2709]: E1124 00:28:41.988784 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:28:46.757243 kubelet[2709]: E1124 00:28:46.757185 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:28:47.758968 kubelet[2709]: E1124 00:28:47.758379 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:28:47.758968 kubelet[2709]: E1124 00:28:47.758861 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:28:47.762580 kubelet[2709]: E1124 00:28:47.760799 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:28:48.757748 kubelet[2709]: E1124 00:28:48.757318 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:28:50.759112 containerd[1559]: time="2025-11-24T00:28:50.759058845Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 24 00:28:50.890750 containerd[1559]: time="2025-11-24T00:28:50.890605640Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:50.891686 containerd[1559]: time="2025-11-24T00:28:50.891587657Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 24 00:28:50.892173 containerd[1559]: time="2025-11-24T00:28:50.892157513Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 24 00:28:50.892799 kubelet[2709]: E1124 00:28:50.892371 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:28:50.892799 kubelet[2709]: E1124 00:28:50.892416 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:28:50.892799 kubelet[2709]: E1124 00:28:50.892521 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:4781a789a9cf40d795f128549fb7a9fc,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2q9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-67b5f8fb49-hhhrv_calico-system(64ccd4c8-759d-46f3-b211-0873db45cfc8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:50.894794 containerd[1559]: time="2025-11-24T00:28:50.894770032Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 24 00:28:51.022791 containerd[1559]: time="2025-11-24T00:28:51.022641398Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:51.023498 containerd[1559]: time="2025-11-24T00:28:51.023435531Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 24 00:28:51.023702 containerd[1559]: time="2025-11-24T00:28:51.023553949Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 24 00:28:51.023746 kubelet[2709]: E1124 00:28:51.023705 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:28:51.023816 kubelet[2709]: E1124 00:28:51.023754 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:28:51.024216 kubelet[2709]: E1124 00:28:51.023863 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m2q9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-67b5f8fb49-hhhrv_calico-system(64ccd4c8-759d-46f3-b211-0873db45cfc8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:51.025940 kubelet[2709]: E1124 00:28:51.025480 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:28:51.756280 kubelet[2709]: E1124 00:28:51.756009 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:28:59.758247 containerd[1559]: time="2025-11-24T00:28:59.758102154Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:28:59.887023 containerd[1559]: time="2025-11-24T00:28:59.886831668Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:28:59.888770 containerd[1559]: time="2025-11-24T00:28:59.887908644Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:28:59.888996 containerd[1559]: time="2025-11-24T00:28:59.888891231Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:28:59.889581 kubelet[2709]: E1124 00:28:59.889508 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:59.889953 kubelet[2709]: E1124 00:28:59.889613 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:28:59.890170 kubelet[2709]: E1124 00:28:59.890084 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h7tpq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-844c564bfd-tnmdj_calico-apiserver(c6030a3d-68f0-4144-8d41-801a65d0dbd4): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:28:59.891308 kubelet[2709]: E1124 00:28:59.891243 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:29:00.760572 containerd[1559]: time="2025-11-24T00:29:00.760330181Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 24 00:29:00.920950 containerd[1559]: time="2025-11-24T00:29:00.920875878Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:00.922161 containerd[1559]: time="2025-11-24T00:29:00.922109933Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 24 00:29:00.922161 containerd[1559]: time="2025-11-24T00:29:00.922188092Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 24 00:29:00.922640 kubelet[2709]: E1124 00:29:00.922562 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:29:00.922640 kubelet[2709]: E1124 00:29:00.922615 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:29:00.924025 containerd[1559]: time="2025-11-24T00:29:00.923046492Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 24 00:29:00.924465 kubelet[2709]: E1124 00:29:00.924426 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:01.058047 containerd[1559]: time="2025-11-24T00:29:01.057384109Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:01.058690 containerd[1559]: time="2025-11-24T00:29:01.058643885Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 24 00:29:01.058788 containerd[1559]: time="2025-11-24T00:29:01.058693924Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 24 00:29:01.058858 kubelet[2709]: E1124 00:29:01.058827 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:29:01.058982 kubelet[2709]: E1124 00:29:01.058926 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:29:01.059425 kubelet[2709]: E1124 00:29:01.059337 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m87kj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-g9zxr_calico-system(be91967a-4242-4c33-899c-c3fa81f6e4dc): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:01.059796 containerd[1559]: time="2025-11-24T00:29:01.059714482Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 24 00:29:01.062031 kubelet[2709]: E1124 00:29:01.061975 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:29:01.194769 containerd[1559]: time="2025-11-24T00:29:01.194689736Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:01.195800 containerd[1559]: time="2025-11-24T00:29:01.195694604Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 24 00:29:01.196089 kubelet[2709]: E1124 00:29:01.195955 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:29:01.196089 kubelet[2709]: E1124 00:29:01.196004 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:29:01.196750 containerd[1559]: time="2025-11-24T00:29:01.195743103Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 24 00:29:01.196785 kubelet[2709]: E1124 00:29:01.196112 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:01.197398 kubelet[2709]: E1124 00:29:01.197325 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:29:02.758531 containerd[1559]: time="2025-11-24T00:29:02.758484131Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 24 00:29:02.885846 containerd[1559]: time="2025-11-24T00:29:02.885785535Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:02.886709 containerd[1559]: time="2025-11-24T00:29:02.886667655Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 24 00:29:02.886754 containerd[1559]: time="2025-11-24T00:29:02.886739284Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 24 00:29:02.886897 kubelet[2709]: E1124 00:29:02.886860 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:29:02.887378 kubelet[2709]: E1124 00:29:02.886907 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:29:02.887378 kubelet[2709]: E1124 00:29:02.887177 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5dwz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5b8f657c8-qdb6r_calico-system(7208ef7a-cd09-4ec2-a692-f04c8a143f91): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:02.887795 containerd[1559]: time="2025-11-24T00:29:02.887765104Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:29:02.888736 kubelet[2709]: E1124 00:29:02.888704 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:29:03.033875 containerd[1559]: time="2025-11-24T00:29:03.032886070Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:03.034019 containerd[1559]: time="2025-11-24T00:29:03.033851971Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:29:03.034019 containerd[1559]: time="2025-11-24T00:29:03.033974260Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:29:03.034439 kubelet[2709]: E1124 00:29:03.034169 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:03.034439 kubelet[2709]: E1124 00:29:03.034265 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:03.034439 kubelet[2709]: E1124 00:29:03.034383 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-69sxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7595d974c5-lnnqz_calico-apiserver(96698920-d098-4174-b103-f535ce1caea1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:03.035831 kubelet[2709]: E1124 00:29:03.035772 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:29:03.755577 kubelet[2709]: E1124 00:29:03.755308 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:29:03.757182 containerd[1559]: time="2025-11-24T00:29:03.757138132Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:29:03.905373 containerd[1559]: time="2025-11-24T00:29:03.905288538Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:03.906574 containerd[1559]: time="2025-11-24T00:29:03.906457477Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:29:03.906574 containerd[1559]: time="2025-11-24T00:29:03.906515176Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:29:03.906985 kubelet[2709]: E1124 00:29:03.906905 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:03.906985 kubelet[2709]: E1124 00:29:03.906991 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:03.907512 kubelet[2709]: E1124 00:29:03.907144 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vrxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-844c564bfd-cldcl_calico-apiserver(d5afe1af-07b4-4767-ad04-88dc81c2e5e8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:03.908817 kubelet[2709]: E1124 00:29:03.908761 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:29:04.760483 kubelet[2709]: E1124 00:29:04.760428 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:29:06.756530 kubelet[2709]: E1124 00:29:06.756483 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:29:09.756140 kubelet[2709]: E1124 00:29:09.756092 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:29:11.756108 kubelet[2709]: E1124 00:29:11.756041 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:29:12.762444 kubelet[2709]: E1124 00:29:12.762400 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:29:14.757632 kubelet[2709]: E1124 00:29:14.757447 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:29:14.757632 kubelet[2709]: E1124 00:29:14.757497 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:29:14.758494 kubelet[2709]: E1124 00:29:14.758276 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:29:15.755906 kubelet[2709]: E1124 00:29:15.755781 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:29:16.758494 kubelet[2709]: E1124 00:29:16.758024 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:29:17.758444 kubelet[2709]: E1124 00:29:17.758368 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:29:24.761936 kubelet[2709]: E1124 00:29:24.761812 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:29:24.763889 kubelet[2709]: E1124 00:29:24.763803 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:29:25.757287 kubelet[2709]: E1124 00:29:25.756053 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:29:26.757195 kubelet[2709]: E1124 00:29:26.756753 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:29:26.758652 kubelet[2709]: E1124 00:29:26.757703 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:29:26.760034 kubelet[2709]: E1124 00:29:26.759700 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:29:28.757054 kubelet[2709]: E1124 00:29:28.756949 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:29:29.756769 kubelet[2709]: E1124 00:29:29.756601 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:29:29.757272 kubelet[2709]: E1124 00:29:29.757201 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:29:30.760327 kubelet[2709]: E1124 00:29:30.760069 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:29:37.756475 kubelet[2709]: E1124 00:29:37.756423 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:29:39.756802 containerd[1559]: time="2025-11-24T00:29:39.756531877Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 24 00:29:39.961629 containerd[1559]: time="2025-11-24T00:29:39.961512524Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:39.962761 containerd[1559]: time="2025-11-24T00:29:39.962663333Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 24 00:29:39.962761 containerd[1559]: time="2025-11-24T00:29:39.962736123Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 24 00:29:39.964050 kubelet[2709]: E1124 00:29:39.963040 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:29:39.964050 kubelet[2709]: E1124 00:29:39.963092 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:29:39.964050 kubelet[2709]: E1124 00:29:39.963199 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:4781a789a9cf40d795f128549fb7a9fc,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2q9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-67b5f8fb49-hhhrv_calico-system(64ccd4c8-759d-46f3-b211-0873db45cfc8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:39.965428 containerd[1559]: time="2025-11-24T00:29:39.965409238Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 24 00:29:40.096445 containerd[1559]: time="2025-11-24T00:29:40.096029348Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:40.097326 containerd[1559]: time="2025-11-24T00:29:40.097297077Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 24 00:29:40.097383 containerd[1559]: time="2025-11-24T00:29:40.097372097Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 24 00:29:40.097691 kubelet[2709]: E1124 00:29:40.097657 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:29:40.097743 kubelet[2709]: E1124 00:29:40.097708 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:29:40.098789 kubelet[2709]: E1124 00:29:40.098749 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m2q9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-67b5f8fb49-hhhrv_calico-system(64ccd4c8-759d-46f3-b211-0873db45cfc8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:40.099949 kubelet[2709]: E1124 00:29:40.099899 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:29:40.755850 kubelet[2709]: E1124 00:29:40.755798 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:29:41.756239 containerd[1559]: time="2025-11-24T00:29:41.756176047Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:29:41.877565 containerd[1559]: time="2025-11-24T00:29:41.877286211Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:41.878402 containerd[1559]: time="2025-11-24T00:29:41.878370412Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:29:41.878460 containerd[1559]: time="2025-11-24T00:29:41.878434962Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:29:41.878676 kubelet[2709]: E1124 00:29:41.878631 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:41.879157 kubelet[2709]: E1124 00:29:41.878675 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:41.879157 kubelet[2709]: E1124 00:29:41.878791 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h7tpq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-844c564bfd-tnmdj_calico-apiserver(c6030a3d-68f0-4144-8d41-801a65d0dbd4): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:41.880044 kubelet[2709]: E1124 00:29:41.879978 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:29:42.757456 containerd[1559]: time="2025-11-24T00:29:42.757400064Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 24 00:29:42.759396 kubelet[2709]: E1124 00:29:42.757880 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:29:42.884684 containerd[1559]: time="2025-11-24T00:29:42.884622706Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:42.885840 containerd[1559]: time="2025-11-24T00:29:42.885794836Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 24 00:29:42.885894 containerd[1559]: time="2025-11-24T00:29:42.885851996Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 24 00:29:42.886033 kubelet[2709]: E1124 00:29:42.885988 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:29:42.886671 kubelet[2709]: E1124 00:29:42.886038 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:29:42.886671 kubelet[2709]: E1124 00:29:42.886126 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:42.888573 containerd[1559]: time="2025-11-24T00:29:42.888548173Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 24 00:29:43.023492 containerd[1559]: time="2025-11-24T00:29:43.023091619Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:43.024728 containerd[1559]: time="2025-11-24T00:29:43.024579536Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 24 00:29:43.024728 containerd[1559]: time="2025-11-24T00:29:43.024702785Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 24 00:29:43.025547 kubelet[2709]: E1124 00:29:43.025147 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:29:43.025673 kubelet[2709]: E1124 00:29:43.025642 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:29:43.027054 kubelet[2709]: E1124 00:29:43.027006 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5vjmh_calico-system(d80649f2-a0ab-46f8-9106-ed888d934bff): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:43.028366 kubelet[2709]: E1124 00:29:43.028316 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:29:44.760526 containerd[1559]: time="2025-11-24T00:29:44.760170517Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:29:44.894197 containerd[1559]: time="2025-11-24T00:29:44.893650887Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:44.895036 containerd[1559]: time="2025-11-24T00:29:44.894990357Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:29:44.895902 containerd[1559]: time="2025-11-24T00:29:44.895788330Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:29:44.896775 kubelet[2709]: E1124 00:29:44.896209 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:44.896775 kubelet[2709]: E1124 00:29:44.896282 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:44.896775 kubelet[2709]: E1124 00:29:44.896413 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vrxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-844c564bfd-cldcl_calico-apiserver(d5afe1af-07b4-4767-ad04-88dc81c2e5e8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:44.898179 kubelet[2709]: E1124 00:29:44.898090 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:29:52.757592 kubelet[2709]: E1124 00:29:52.757515 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:29:52.759107 kubelet[2709]: E1124 00:29:52.758684 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:29:52.759162 containerd[1559]: time="2025-11-24T00:29:52.757725343Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:29:52.878064 containerd[1559]: time="2025-11-24T00:29:52.877838069Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:52.879161 containerd[1559]: time="2025-11-24T00:29:52.879136640Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:29:52.879292 containerd[1559]: time="2025-11-24T00:29:52.879180840Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:29:52.880225 kubelet[2709]: E1124 00:29:52.880030 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:52.880225 kubelet[2709]: E1124 00:29:52.880087 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:29:52.881010 kubelet[2709]: E1124 00:29:52.880472 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-69sxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-7595d974c5-lnnqz_calico-apiserver(96698920-d098-4174-b103-f535ce1caea1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:52.881137 containerd[1559]: time="2025-11-24T00:29:52.880710210Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 24 00:29:52.881605 kubelet[2709]: E1124 00:29:52.881581 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:29:53.006515 containerd[1559]: time="2025-11-24T00:29:53.006156264Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:53.007408 containerd[1559]: time="2025-11-24T00:29:53.007184327Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 24 00:29:53.008088 containerd[1559]: time="2025-11-24T00:29:53.007492535Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 24 00:29:53.008135 kubelet[2709]: E1124 00:29:53.007698 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:29:53.008135 kubelet[2709]: E1124 00:29:53.007746 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:29:53.008135 kubelet[2709]: E1124 00:29:53.007856 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m87kj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-g9zxr_calico-system(be91967a-4242-4c33-899c-c3fa81f6e4dc): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:53.009983 kubelet[2709]: E1124 00:29:53.009956 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:29:53.756876 kubelet[2709]: E1124 00:29:53.756786 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:29:57.757371 containerd[1559]: time="2025-11-24T00:29:57.757293705Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 24 00:29:57.759860 kubelet[2709]: E1124 00:29:57.759103 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:29:57.883774 containerd[1559]: time="2025-11-24T00:29:57.883685832Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:29:57.885056 containerd[1559]: time="2025-11-24T00:29:57.885010315Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 24 00:29:57.885135 containerd[1559]: time="2025-11-24T00:29:57.885105945Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 24 00:29:57.885539 kubelet[2709]: E1124 00:29:57.885490 2709 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:29:57.885610 kubelet[2709]: E1124 00:29:57.885546 2709 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:29:57.885710 kubelet[2709]: E1124 00:29:57.885655 2709 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5dwz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5b8f657c8-qdb6r_calico-system(7208ef7a-cd09-4ec2-a692-f04c8a143f91): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 24 00:29:57.887193 kubelet[2709]: E1124 00:29:57.887141 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:30:03.337071 systemd[1]: Started sshd@7-172.237.154.236:22-139.178.68.195:40270.service - OpenSSH per-connection server daemon (139.178.68.195:40270). Nov 24 00:30:03.714618 sshd[5001]: Accepted publickey for core from 139.178.68.195 port 40270 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:03.716819 sshd-session[5001]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:03.722985 systemd-logind[1525]: New session 8 of user core. Nov 24 00:30:03.729129 systemd[1]: Started session-8.scope - Session 8 of User core. Nov 24 00:30:04.064991 sshd[5004]: Connection closed by 139.178.68.195 port 40270 Nov 24 00:30:04.064873 sshd-session[5001]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:04.074177 systemd[1]: sshd@7-172.237.154.236:22-139.178.68.195:40270.service: Deactivated successfully. Nov 24 00:30:04.079875 systemd[1]: session-8.scope: Deactivated successfully. Nov 24 00:30:04.082879 systemd-logind[1525]: Session 8 logged out. Waiting for processes to exit. Nov 24 00:30:04.085640 systemd-logind[1525]: Removed session 8. Nov 24 00:30:04.757817 kubelet[2709]: E1124 00:30:04.757715 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:30:05.760939 kubelet[2709]: E1124 00:30:05.760707 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:30:05.760939 kubelet[2709]: E1124 00:30:05.760774 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:30:05.762354 kubelet[2709]: E1124 00:30:05.762311 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:30:05.762636 kubelet[2709]: E1124 00:30:05.762520 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:30:09.127775 systemd[1]: Started sshd@8-172.237.154.236:22-139.178.68.195:40274.service - OpenSSH per-connection server daemon (139.178.68.195:40274). Nov 24 00:30:09.491601 sshd[5021]: Accepted publickey for core from 139.178.68.195 port 40274 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:09.493429 sshd-session[5021]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:09.500354 systemd-logind[1525]: New session 9 of user core. Nov 24 00:30:09.507027 systemd[1]: Started session-9.scope - Session 9 of User core. Nov 24 00:30:09.816161 sshd[5024]: Connection closed by 139.178.68.195 port 40274 Nov 24 00:30:09.818224 sshd-session[5021]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:09.823285 systemd-logind[1525]: Session 9 logged out. Waiting for processes to exit. Nov 24 00:30:09.824417 systemd[1]: sshd@8-172.237.154.236:22-139.178.68.195:40274.service: Deactivated successfully. Nov 24 00:30:09.829287 systemd[1]: session-9.scope: Deactivated successfully. Nov 24 00:30:09.832812 systemd-logind[1525]: Removed session 9. Nov 24 00:30:10.762135 kubelet[2709]: E1124 00:30:10.761352 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:30:10.764061 kubelet[2709]: E1124 00:30:10.762810 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:30:12.760799 kubelet[2709]: E1124 00:30:12.760723 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:30:14.755170 kubelet[2709]: E1124 00:30:14.755101 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:30:14.873560 systemd[1]: Started sshd@9-172.237.154.236:22-139.178.68.195:46614.service - OpenSSH per-connection server daemon (139.178.68.195:46614). Nov 24 00:30:15.218723 sshd[5061]: Accepted publickey for core from 139.178.68.195 port 46614 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:15.221885 sshd-session[5061]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:15.233408 systemd-logind[1525]: New session 10 of user core. Nov 24 00:30:15.240462 systemd[1]: Started session-10.scope - Session 10 of User core. Nov 24 00:30:15.544433 sshd[5064]: Connection closed by 139.178.68.195 port 46614 Nov 24 00:30:15.547389 sshd-session[5061]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:15.553197 systemd[1]: sshd@9-172.237.154.236:22-139.178.68.195:46614.service: Deactivated successfully. Nov 24 00:30:15.556342 systemd[1]: session-10.scope: Deactivated successfully. Nov 24 00:30:15.559836 systemd-logind[1525]: Session 10 logged out. Waiting for processes to exit. Nov 24 00:30:15.565716 systemd-logind[1525]: Removed session 10. Nov 24 00:30:15.613226 systemd[1]: Started sshd@10-172.237.154.236:22-139.178.68.195:46618.service - OpenSSH per-connection server daemon (139.178.68.195:46618). Nov 24 00:30:15.976982 sshd[5077]: Accepted publickey for core from 139.178.68.195 port 46618 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:15.979325 sshd-session[5077]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:15.992617 systemd-logind[1525]: New session 11 of user core. Nov 24 00:30:15.997382 systemd[1]: Started session-11.scope - Session 11 of User core. Nov 24 00:30:16.347617 sshd[5083]: Connection closed by 139.178.68.195 port 46618 Nov 24 00:30:16.348168 sshd-session[5077]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:16.356531 systemd-logind[1525]: Session 11 logged out. Waiting for processes to exit. Nov 24 00:30:16.358438 systemd[1]: sshd@10-172.237.154.236:22-139.178.68.195:46618.service: Deactivated successfully. Nov 24 00:30:16.361280 systemd[1]: session-11.scope: Deactivated successfully. Nov 24 00:30:16.364186 systemd-logind[1525]: Removed session 11. Nov 24 00:30:16.410686 systemd[1]: Started sshd@11-172.237.154.236:22-139.178.68.195:46624.service - OpenSSH per-connection server daemon (139.178.68.195:46624). Nov 24 00:30:16.761051 kubelet[2709]: E1124 00:30:16.760998 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:30:16.765129 kubelet[2709]: E1124 00:30:16.765084 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:30:16.782494 sshd[5094]: Accepted publickey for core from 139.178.68.195 port 46624 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:16.784608 sshd-session[5094]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:16.794146 systemd-logind[1525]: New session 12 of user core. Nov 24 00:30:16.800758 systemd[1]: Started session-12.scope - Session 12 of User core. Nov 24 00:30:17.113175 sshd[5097]: Connection closed by 139.178.68.195 port 46624 Nov 24 00:30:17.114139 sshd-session[5094]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:17.122646 systemd[1]: sshd@11-172.237.154.236:22-139.178.68.195:46624.service: Deactivated successfully. Nov 24 00:30:17.123050 systemd-logind[1525]: Session 12 logged out. Waiting for processes to exit. Nov 24 00:30:17.129094 systemd[1]: session-12.scope: Deactivated successfully. Nov 24 00:30:17.135432 systemd-logind[1525]: Removed session 12. Nov 24 00:30:17.756708 kubelet[2709]: E1124 00:30:17.756657 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:30:17.757393 kubelet[2709]: E1124 00:30:17.757371 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:30:17.757589 kubelet[2709]: E1124 00:30:17.757428 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:30:18.757901 kubelet[2709]: E1124 00:30:18.757863 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:30:22.177990 systemd[1]: Started sshd@12-172.237.154.236:22-139.178.68.195:32850.service - OpenSSH per-connection server daemon (139.178.68.195:32850). Nov 24 00:30:22.517801 sshd[5111]: Accepted publickey for core from 139.178.68.195 port 32850 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:22.520530 sshd-session[5111]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:22.527971 systemd-logind[1525]: New session 13 of user core. Nov 24 00:30:22.538081 systemd[1]: Started session-13.scope - Session 13 of User core. Nov 24 00:30:22.757603 kubelet[2709]: E1124 00:30:22.757560 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:30:22.821711 sshd[5114]: Connection closed by 139.178.68.195 port 32850 Nov 24 00:30:22.822614 sshd-session[5111]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:22.826951 systemd-logind[1525]: Session 13 logged out. Waiting for processes to exit. Nov 24 00:30:22.827727 systemd[1]: sshd@12-172.237.154.236:22-139.178.68.195:32850.service: Deactivated successfully. Nov 24 00:30:22.830204 systemd[1]: session-13.scope: Deactivated successfully. Nov 24 00:30:22.832145 systemd-logind[1525]: Removed session 13. Nov 24 00:30:22.886276 systemd[1]: Started sshd@13-172.237.154.236:22-139.178.68.195:32864.service - OpenSSH per-connection server daemon (139.178.68.195:32864). Nov 24 00:30:23.234235 sshd[5125]: Accepted publickey for core from 139.178.68.195 port 32864 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:23.236308 sshd-session[5125]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:23.248730 systemd-logind[1525]: New session 14 of user core. Nov 24 00:30:23.253322 systemd[1]: Started session-14.scope - Session 14 of User core. Nov 24 00:30:23.758228 kubelet[2709]: E1124 00:30:23.757592 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:30:23.960607 sshd[5128]: Connection closed by 139.178.68.195 port 32864 Nov 24 00:30:23.962340 sshd-session[5125]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:23.967142 systemd[1]: sshd@13-172.237.154.236:22-139.178.68.195:32864.service: Deactivated successfully. Nov 24 00:30:23.969409 systemd[1]: session-14.scope: Deactivated successfully. Nov 24 00:30:23.970754 systemd-logind[1525]: Session 14 logged out. Waiting for processes to exit. Nov 24 00:30:23.972582 systemd-logind[1525]: Removed session 14. Nov 24 00:30:24.023945 systemd[1]: Started sshd@14-172.237.154.236:22-139.178.68.195:32880.service - OpenSSH per-connection server daemon (139.178.68.195:32880). Nov 24 00:30:24.362083 sshd[5138]: Accepted publickey for core from 139.178.68.195 port 32880 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:24.363791 sshd-session[5138]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:24.369830 systemd-logind[1525]: New session 15 of user core. Nov 24 00:30:24.380064 systemd[1]: Started session-15.scope - Session 15 of User core. Nov 24 00:30:25.315773 sshd[5141]: Connection closed by 139.178.68.195 port 32880 Nov 24 00:30:25.316438 sshd-session[5138]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:25.321156 systemd-logind[1525]: Session 15 logged out. Waiting for processes to exit. Nov 24 00:30:25.322142 systemd[1]: sshd@14-172.237.154.236:22-139.178.68.195:32880.service: Deactivated successfully. Nov 24 00:30:25.326866 systemd[1]: session-15.scope: Deactivated successfully. Nov 24 00:30:25.332587 systemd-logind[1525]: Removed session 15. Nov 24 00:30:25.379347 systemd[1]: Started sshd@15-172.237.154.236:22-139.178.68.195:32892.service - OpenSSH per-connection server daemon (139.178.68.195:32892). Nov 24 00:30:25.720083 sshd[5158]: Accepted publickey for core from 139.178.68.195 port 32892 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:25.722855 sshd-session[5158]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:25.731597 systemd-logind[1525]: New session 16 of user core. Nov 24 00:30:25.739088 systemd[1]: Started session-16.scope - Session 16 of User core. Nov 24 00:30:26.124292 sshd[5161]: Connection closed by 139.178.68.195 port 32892 Nov 24 00:30:26.125107 sshd-session[5158]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:26.130123 systemd[1]: sshd@15-172.237.154.236:22-139.178.68.195:32892.service: Deactivated successfully. Nov 24 00:30:26.134006 systemd[1]: session-16.scope: Deactivated successfully. Nov 24 00:30:26.135862 systemd-logind[1525]: Session 16 logged out. Waiting for processes to exit. Nov 24 00:30:26.137648 systemd-logind[1525]: Removed session 16. Nov 24 00:30:26.195130 systemd[1]: Started sshd@16-172.237.154.236:22-139.178.68.195:32894.service - OpenSSH per-connection server daemon (139.178.68.195:32894). Nov 24 00:30:26.562889 sshd[5171]: Accepted publickey for core from 139.178.68.195 port 32894 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:26.565432 sshd-session[5171]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:26.572552 systemd-logind[1525]: New session 17 of user core. Nov 24 00:30:26.578141 systemd[1]: Started session-17.scope - Session 17 of User core. Nov 24 00:30:26.904559 sshd[5174]: Connection closed by 139.178.68.195 port 32894 Nov 24 00:30:26.905660 sshd-session[5171]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:26.913965 systemd[1]: sshd@16-172.237.154.236:22-139.178.68.195:32894.service: Deactivated successfully. Nov 24 00:30:26.914847 systemd-logind[1525]: Session 17 logged out. Waiting for processes to exit. Nov 24 00:30:26.918158 systemd[1]: session-17.scope: Deactivated successfully. Nov 24 00:30:26.924060 systemd-logind[1525]: Removed session 17. Nov 24 00:30:28.760571 kubelet[2709]: E1124 00:30:28.760498 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:30:29.756598 kubelet[2709]: E1124 00:30:29.756445 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:30:29.763158 kubelet[2709]: E1124 00:30:29.762808 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:30:29.763158 kubelet[2709]: E1124 00:30:29.762808 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:30:30.756044 kubelet[2709]: E1124 00:30:30.756003 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc" Nov 24 00:30:31.962559 systemd[1]: Started sshd@17-172.237.154.236:22-139.178.68.195:43160.service - OpenSSH per-connection server daemon (139.178.68.195:43160). Nov 24 00:30:32.300230 sshd[5188]: Accepted publickey for core from 139.178.68.195 port 43160 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:32.302240 sshd-session[5188]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:32.307561 systemd-logind[1525]: New session 18 of user core. Nov 24 00:30:32.315110 systemd[1]: Started session-18.scope - Session 18 of User core. Nov 24 00:30:32.611213 sshd[5191]: Connection closed by 139.178.68.195 port 43160 Nov 24 00:30:32.613176 sshd-session[5188]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:32.618942 systemd-logind[1525]: Session 18 logged out. Waiting for processes to exit. Nov 24 00:30:32.621493 systemd[1]: sshd@17-172.237.154.236:22-139.178.68.195:43160.service: Deactivated successfully. Nov 24 00:30:32.627644 systemd[1]: session-18.scope: Deactivated successfully. Nov 24 00:30:32.630323 systemd-logind[1525]: Removed session 18. Nov 24 00:30:32.759205 kubelet[2709]: E1124 00:30:32.759146 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5vjmh" podUID="d80649f2-a0ab-46f8-9106-ed888d934bff" Nov 24 00:30:34.755210 kubelet[2709]: E1124 00:30:34.755155 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:30:35.755461 kubelet[2709]: E1124 00:30:35.755381 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:30:35.757955 kubelet[2709]: E1124 00:30:35.756449 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-cldcl" podUID="d5afe1af-07b4-4767-ad04-88dc81c2e5e8" Nov 24 00:30:36.757944 kubelet[2709]: E1124 00:30:36.755826 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:30:37.681241 systemd[1]: Started sshd@18-172.237.154.236:22-139.178.68.195:43172.service - OpenSSH per-connection server daemon (139.178.68.195:43172). Nov 24 00:30:37.756657 kubelet[2709]: E1124 00:30:37.756619 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5b8f657c8-qdb6r" podUID="7208ef7a-cd09-4ec2-a692-f04c8a143f91" Nov 24 00:30:38.042284 sshd[5204]: Accepted publickey for core from 139.178.68.195 port 43172 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:38.044362 sshd-session[5204]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:38.052021 systemd-logind[1525]: New session 19 of user core. Nov 24 00:30:38.059202 systemd[1]: Started session-19.scope - Session 19 of User core. Nov 24 00:30:38.377963 sshd[5207]: Connection closed by 139.178.68.195 port 43172 Nov 24 00:30:38.378952 sshd-session[5204]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:38.384986 systemd-logind[1525]: Session 19 logged out. Waiting for processes to exit. Nov 24 00:30:38.388355 systemd[1]: sshd@18-172.237.154.236:22-139.178.68.195:43172.service: Deactivated successfully. Nov 24 00:30:38.395649 systemd[1]: session-19.scope: Deactivated successfully. Nov 24 00:30:38.400340 systemd-logind[1525]: Removed session 19. Nov 24 00:30:39.756202 kubelet[2709]: E1124 00:30:39.755786 2709 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.19 172.232.0.20 172.232.0.15" Nov 24 00:30:39.759289 kubelet[2709]: E1124 00:30:39.759261 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-67b5f8fb49-hhhrv" podUID="64ccd4c8-759d-46f3-b211-0873db45cfc8" Nov 24 00:30:41.756549 kubelet[2709]: E1124 00:30:41.756493 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-844c564bfd-tnmdj" podUID="c6030a3d-68f0-4144-8d41-801a65d0dbd4" Nov 24 00:30:43.448235 systemd[1]: Started sshd@19-172.237.154.236:22-139.178.68.195:48478.service - OpenSSH per-connection server daemon (139.178.68.195:48478). Nov 24 00:30:43.799601 sshd[5245]: Accepted publickey for core from 139.178.68.195 port 48478 ssh2: RSA SHA256:Z+qT+FuO0vejOh7g8jz95x+2xvOWEpZ6V336WAQdLnE Nov 24 00:30:43.802079 sshd-session[5245]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:30:43.807648 systemd-logind[1525]: New session 20 of user core. Nov 24 00:30:43.818039 systemd[1]: Started session-20.scope - Session 20 of User core. Nov 24 00:30:44.156105 sshd[5248]: Connection closed by 139.178.68.195 port 48478 Nov 24 00:30:44.159085 sshd-session[5245]: pam_unix(sshd:session): session closed for user core Nov 24 00:30:44.163834 systemd-logind[1525]: Session 20 logged out. Waiting for processes to exit. Nov 24 00:30:44.164861 systemd[1]: sshd@19-172.237.154.236:22-139.178.68.195:48478.service: Deactivated successfully. Nov 24 00:30:44.170988 systemd[1]: session-20.scope: Deactivated successfully. Nov 24 00:30:44.176179 systemd-logind[1525]: Removed session 20. Nov 24 00:30:44.757113 kubelet[2709]: E1124 00:30:44.757028 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-7595d974c5-lnnqz" podUID="96698920-d098-4174-b103-f535ce1caea1" Nov 24 00:30:45.756596 kubelet[2709]: E1124 00:30:45.755824 2709 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-g9zxr" podUID="be91967a-4242-4c33-899c-c3fa81f6e4dc"