Nov 24 00:13:53.088175 kernel: Linux version 6.12.58-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Sun Nov 23 20:54:38 -00 2025 Nov 24 00:13:53.088207 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=1969a6ee0c0ec5507eb68849c160e94c58e52d2291c767873af68a1f52b30801 Nov 24 00:13:53.088221 kernel: BIOS-provided physical RAM map: Nov 24 00:13:53.088230 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Nov 24 00:13:53.088238 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Nov 24 00:13:53.088247 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Nov 24 00:13:53.088257 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Nov 24 00:13:53.088266 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Nov 24 00:13:53.088279 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Nov 24 00:13:53.088288 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Nov 24 00:13:53.088297 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Nov 24 00:13:53.088309 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Nov 24 00:13:53.088318 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Nov 24 00:13:53.088327 kernel: NX (Execute Disable) protection: active Nov 24 00:13:53.088338 kernel: APIC: Static calls initialized Nov 24 00:13:53.088347 kernel: SMBIOS 2.8 present. Nov 24 00:13:53.088364 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Nov 24 00:13:53.088373 kernel: DMI: Memory slots populated: 1/1 Nov 24 00:13:53.088382 kernel: Hypervisor detected: KVM Nov 24 00:13:53.088392 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Nov 24 00:13:53.088401 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Nov 24 00:13:53.088411 kernel: kvm-clock: using sched offset of 5135576799 cycles Nov 24 00:13:53.088420 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Nov 24 00:13:53.088430 kernel: tsc: Detected 2794.748 MHz processor Nov 24 00:13:53.088440 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Nov 24 00:13:53.088450 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Nov 24 00:13:53.088464 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Nov 24 00:13:53.088474 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Nov 24 00:13:53.088484 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Nov 24 00:13:53.088494 kernel: Using GB pages for direct mapping Nov 24 00:13:53.088504 kernel: ACPI: Early table checksum verification disabled Nov 24 00:13:53.088514 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Nov 24 00:13:53.088524 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:13:53.088534 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:13:53.088544 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:13:53.088558 kernel: ACPI: FACS 0x000000009CFE0000 000040 Nov 24 00:13:53.088568 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:13:53.088578 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:13:53.088588 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:13:53.088599 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 24 00:13:53.088614 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Nov 24 00:13:53.088624 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Nov 24 00:13:53.088638 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Nov 24 00:13:53.088648 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Nov 24 00:13:53.088658 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Nov 24 00:13:53.088669 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Nov 24 00:13:53.088679 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Nov 24 00:13:53.088689 kernel: No NUMA configuration found Nov 24 00:13:53.088700 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Nov 24 00:13:53.088713 kernel: NODE_DATA(0) allocated [mem 0x9cfd4dc0-0x9cfdbfff] Nov 24 00:13:53.088724 kernel: Zone ranges: Nov 24 00:13:53.088734 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Nov 24 00:13:53.088745 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Nov 24 00:13:53.088755 kernel: Normal empty Nov 24 00:13:53.088766 kernel: Device empty Nov 24 00:13:53.088778 kernel: Movable zone start for each node Nov 24 00:13:53.088791 kernel: Early memory node ranges Nov 24 00:13:53.088804 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Nov 24 00:13:53.088816 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Nov 24 00:13:53.088833 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Nov 24 00:13:53.088847 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Nov 24 00:13:53.088860 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Nov 24 00:13:53.088873 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Nov 24 00:13:53.088891 kernel: ACPI: PM-Timer IO Port: 0x608 Nov 24 00:13:53.088904 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Nov 24 00:13:53.088917 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Nov 24 00:13:53.088930 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Nov 24 00:13:53.088946 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Nov 24 00:13:53.088964 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Nov 24 00:13:53.088977 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Nov 24 00:13:53.088990 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Nov 24 00:13:53.089001 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Nov 24 00:13:53.089011 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Nov 24 00:13:53.089033 kernel: TSC deadline timer available Nov 24 00:13:53.089044 kernel: CPU topo: Max. logical packages: 1 Nov 24 00:13:53.089069 kernel: CPU topo: Max. logical dies: 1 Nov 24 00:13:53.089080 kernel: CPU topo: Max. dies per package: 1 Nov 24 00:13:53.089090 kernel: CPU topo: Max. threads per core: 1 Nov 24 00:13:53.089106 kernel: CPU topo: Num. cores per package: 4 Nov 24 00:13:53.089116 kernel: CPU topo: Num. threads per package: 4 Nov 24 00:13:53.089126 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Nov 24 00:13:53.089136 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Nov 24 00:13:53.089146 kernel: kvm-guest: KVM setup pv remote TLB flush Nov 24 00:13:53.089155 kernel: kvm-guest: setup PV sched yield Nov 24 00:13:53.089165 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Nov 24 00:13:53.089174 kernel: Booting paravirtualized kernel on KVM Nov 24 00:13:53.089185 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Nov 24 00:13:53.089199 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Nov 24 00:13:53.089210 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Nov 24 00:13:53.089221 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Nov 24 00:13:53.089231 kernel: pcpu-alloc: [0] 0 1 2 3 Nov 24 00:13:53.089241 kernel: kvm-guest: PV spinlocks enabled Nov 24 00:13:53.089251 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Nov 24 00:13:53.089263 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=1969a6ee0c0ec5507eb68849c160e94c58e52d2291c767873af68a1f52b30801 Nov 24 00:13:53.089274 kernel: random: crng init done Nov 24 00:13:53.089288 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Nov 24 00:13:53.089298 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 24 00:13:53.089309 kernel: Fallback order for Node 0: 0 Nov 24 00:13:53.089319 kernel: Built 1 zonelists, mobility grouping on. Total pages: 642938 Nov 24 00:13:53.089330 kernel: Policy zone: DMA32 Nov 24 00:13:53.089340 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Nov 24 00:13:53.089351 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Nov 24 00:13:53.089361 kernel: ftrace: allocating 40103 entries in 157 pages Nov 24 00:13:53.089371 kernel: ftrace: allocated 157 pages with 5 groups Nov 24 00:13:53.089384 kernel: Dynamic Preempt: voluntary Nov 24 00:13:53.089395 kernel: rcu: Preemptible hierarchical RCU implementation. Nov 24 00:13:53.089406 kernel: rcu: RCU event tracing is enabled. Nov 24 00:13:53.089417 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Nov 24 00:13:53.089428 kernel: Trampoline variant of Tasks RCU enabled. Nov 24 00:13:53.089442 kernel: Rude variant of Tasks RCU enabled. Nov 24 00:13:53.089453 kernel: Tracing variant of Tasks RCU enabled. Nov 24 00:13:53.089463 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Nov 24 00:13:53.089474 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Nov 24 00:13:53.089484 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 24 00:13:53.089498 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 24 00:13:53.089508 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 24 00:13:53.089519 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Nov 24 00:13:53.089529 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Nov 24 00:13:53.089549 kernel: Console: colour VGA+ 80x25 Nov 24 00:13:53.089563 kernel: printk: legacy console [ttyS0] enabled Nov 24 00:13:53.089573 kernel: ACPI: Core revision 20240827 Nov 24 00:13:53.089584 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Nov 24 00:13:53.089595 kernel: APIC: Switch to symmetric I/O mode setup Nov 24 00:13:53.089605 kernel: x2apic enabled Nov 24 00:13:53.089616 kernel: APIC: Switched APIC routing to: physical x2apic Nov 24 00:13:53.089635 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Nov 24 00:13:53.089646 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Nov 24 00:13:53.089657 kernel: kvm-guest: setup PV IPIs Nov 24 00:13:53.089668 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Nov 24 00:13:53.089679 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Nov 24 00:13:53.089694 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Nov 24 00:13:53.089705 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Nov 24 00:13:53.089716 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Nov 24 00:13:53.089727 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Nov 24 00:13:53.089738 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Nov 24 00:13:53.089749 kernel: Spectre V2 : Mitigation: Retpolines Nov 24 00:13:53.089759 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Nov 24 00:13:53.089770 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Nov 24 00:13:53.089782 kernel: active return thunk: retbleed_return_thunk Nov 24 00:13:53.089797 kernel: RETBleed: Mitigation: untrained return thunk Nov 24 00:13:53.089808 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Nov 24 00:13:53.089822 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Nov 24 00:13:53.089834 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Nov 24 00:13:53.089847 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Nov 24 00:13:53.089858 kernel: active return thunk: srso_return_thunk Nov 24 00:13:53.089869 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Nov 24 00:13:53.089880 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Nov 24 00:13:53.089894 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Nov 24 00:13:53.089905 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Nov 24 00:13:53.089916 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Nov 24 00:13:53.089927 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Nov 24 00:13:53.089938 kernel: Freeing SMP alternatives memory: 32K Nov 24 00:13:53.089949 kernel: pid_max: default: 32768 minimum: 301 Nov 24 00:13:53.089960 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Nov 24 00:13:53.089971 kernel: landlock: Up and running. Nov 24 00:13:53.089982 kernel: SELinux: Initializing. Nov 24 00:13:53.090001 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 24 00:13:53.090013 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 24 00:13:53.090034 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Nov 24 00:13:53.090046 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Nov 24 00:13:53.090072 kernel: ... version: 0 Nov 24 00:13:53.090084 kernel: ... bit width: 48 Nov 24 00:13:53.090094 kernel: ... generic registers: 6 Nov 24 00:13:53.090105 kernel: ... value mask: 0000ffffffffffff Nov 24 00:13:53.090116 kernel: ... max period: 00007fffffffffff Nov 24 00:13:53.090132 kernel: ... fixed-purpose events: 0 Nov 24 00:13:53.090143 kernel: ... event mask: 000000000000003f Nov 24 00:13:53.090154 kernel: signal: max sigframe size: 1776 Nov 24 00:13:53.090165 kernel: rcu: Hierarchical SRCU implementation. Nov 24 00:13:53.090176 kernel: rcu: Max phase no-delay instances is 400. Nov 24 00:13:53.090187 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Nov 24 00:13:53.090198 kernel: smp: Bringing up secondary CPUs ... Nov 24 00:13:53.090209 kernel: smpboot: x86: Booting SMP configuration: Nov 24 00:13:53.090220 kernel: .... node #0, CPUs: #1 #2 #3 Nov 24 00:13:53.090235 kernel: smp: Brought up 1 node, 4 CPUs Nov 24 00:13:53.090246 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Nov 24 00:13:53.090257 kernel: Memory: 2420720K/2571752K available (14336K kernel code, 2444K rwdata, 26064K rodata, 46188K init, 2572K bss, 145096K reserved, 0K cma-reserved) Nov 24 00:13:53.090268 kernel: devtmpfs: initialized Nov 24 00:13:53.090279 kernel: x86/mm: Memory block size: 128MB Nov 24 00:13:53.090291 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Nov 24 00:13:53.090302 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Nov 24 00:13:53.090313 kernel: pinctrl core: initialized pinctrl subsystem Nov 24 00:13:53.090328 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Nov 24 00:13:53.090340 kernel: audit: initializing netlink subsys (disabled) Nov 24 00:13:53.090350 kernel: audit: type=2000 audit(1763943230.046:1): state=initialized audit_enabled=0 res=1 Nov 24 00:13:53.090360 kernel: thermal_sys: Registered thermal governor 'step_wise' Nov 24 00:13:53.090371 kernel: thermal_sys: Registered thermal governor 'user_space' Nov 24 00:13:53.090382 kernel: cpuidle: using governor menu Nov 24 00:13:53.090392 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Nov 24 00:13:53.090403 kernel: dca service started, version 1.12.1 Nov 24 00:13:53.090414 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] (base 0xb0000000) for domain 0000 [bus 00-ff] Nov 24 00:13:53.090425 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] reserved as E820 entry Nov 24 00:13:53.090439 kernel: PCI: Using configuration type 1 for base access Nov 24 00:13:53.090451 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Nov 24 00:13:53.090462 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Nov 24 00:13:53.090473 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Nov 24 00:13:53.090484 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Nov 24 00:13:53.090494 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Nov 24 00:13:53.090505 kernel: ACPI: Added _OSI(Module Device) Nov 24 00:13:53.090516 kernel: ACPI: Added _OSI(Processor Device) Nov 24 00:13:53.090527 kernel: ACPI: Added _OSI(Processor Aggregator Device) Nov 24 00:13:53.090542 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Nov 24 00:13:53.090557 kernel: ACPI: Interpreter enabled Nov 24 00:13:53.090569 kernel: ACPI: PM: (supports S0 S3 S5) Nov 24 00:13:53.090580 kernel: ACPI: Using IOAPIC for interrupt routing Nov 24 00:13:53.090592 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Nov 24 00:13:53.090603 kernel: PCI: Using E820 reservations for host bridge windows Nov 24 00:13:53.090614 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Nov 24 00:13:53.090625 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Nov 24 00:13:53.090953 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Nov 24 00:13:53.091159 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Nov 24 00:13:53.091324 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Nov 24 00:13:53.091341 kernel: PCI host bridge to bus 0000:00 Nov 24 00:13:53.091509 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Nov 24 00:13:53.091660 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Nov 24 00:13:53.091804 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Nov 24 00:13:53.092007 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Nov 24 00:13:53.092193 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Nov 24 00:13:53.092343 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Nov 24 00:13:53.092492 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Nov 24 00:13:53.092683 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Nov 24 00:13:53.092856 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Nov 24 00:13:53.093019 kernel: pci 0000:00:01.0: BAR 0 [mem 0xfd000000-0xfdffffff pref] Nov 24 00:13:53.093215 kernel: pci 0000:00:01.0: BAR 2 [mem 0xfebd0000-0xfebd0fff] Nov 24 00:13:53.093378 kernel: pci 0000:00:01.0: ROM [mem 0xfebc0000-0xfebcffff pref] Nov 24 00:13:53.093538 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Nov 24 00:13:53.093709 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Nov 24 00:13:53.093872 kernel: pci 0000:00:02.0: BAR 0 [io 0xc0c0-0xc0df] Nov 24 00:13:53.094044 kernel: pci 0000:00:02.0: BAR 1 [mem 0xfebd1000-0xfebd1fff] Nov 24 00:13:53.094230 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfe000000-0xfe003fff 64bit pref] Nov 24 00:13:53.094397 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Nov 24 00:13:53.094548 kernel: pci 0000:00:03.0: BAR 0 [io 0xc000-0xc07f] Nov 24 00:13:53.094699 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebd2000-0xfebd2fff] Nov 24 00:13:53.094853 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe004000-0xfe007fff 64bit pref] Nov 24 00:13:53.095050 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Nov 24 00:13:53.095231 kernel: pci 0000:00:04.0: BAR 0 [io 0xc0e0-0xc0ff] Nov 24 00:13:53.095396 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebd3000-0xfebd3fff] Nov 24 00:13:53.095544 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe008000-0xfe00bfff 64bit pref] Nov 24 00:13:53.095692 kernel: pci 0000:00:04.0: ROM [mem 0xfeb80000-0xfebbffff pref] Nov 24 00:13:53.095847 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Nov 24 00:13:53.096001 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Nov 24 00:13:53.096196 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Nov 24 00:13:53.096348 kernel: pci 0000:00:1f.2: BAR 4 [io 0xc100-0xc11f] Nov 24 00:13:53.096503 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xfebd4000-0xfebd4fff] Nov 24 00:13:53.096675 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Nov 24 00:13:53.096833 kernel: pci 0000:00:1f.3: BAR 4 [io 0x0700-0x073f] Nov 24 00:13:53.096849 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Nov 24 00:13:53.096863 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Nov 24 00:13:53.096876 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Nov 24 00:13:53.096894 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Nov 24 00:13:53.096911 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Nov 24 00:13:53.096922 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Nov 24 00:13:53.096933 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Nov 24 00:13:53.096944 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Nov 24 00:13:53.096955 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Nov 24 00:13:53.096967 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Nov 24 00:13:53.096978 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Nov 24 00:13:53.096989 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Nov 24 00:13:53.097000 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Nov 24 00:13:53.097015 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Nov 24 00:13:53.097037 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Nov 24 00:13:53.097048 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Nov 24 00:13:53.097075 kernel: iommu: Default domain type: Translated Nov 24 00:13:53.097086 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Nov 24 00:13:53.097097 kernel: PCI: Using ACPI for IRQ routing Nov 24 00:13:53.097108 kernel: PCI: pci_cache_line_size set to 64 bytes Nov 24 00:13:53.097119 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Nov 24 00:13:53.097130 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Nov 24 00:13:53.097306 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Nov 24 00:13:53.097472 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Nov 24 00:13:53.097640 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Nov 24 00:13:53.097657 kernel: vgaarb: loaded Nov 24 00:13:53.097669 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Nov 24 00:13:53.097680 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Nov 24 00:13:53.097691 kernel: clocksource: Switched to clocksource kvm-clock Nov 24 00:13:53.097703 kernel: VFS: Disk quotas dquot_6.6.0 Nov 24 00:13:53.097714 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Nov 24 00:13:53.097730 kernel: pnp: PnP ACPI init Nov 24 00:13:53.097899 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Nov 24 00:13:53.097916 kernel: pnp: PnP ACPI: found 6 devices Nov 24 00:13:53.097928 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Nov 24 00:13:53.097940 kernel: NET: Registered PF_INET protocol family Nov 24 00:13:53.097952 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 24 00:13:53.097963 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Nov 24 00:13:53.097974 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Nov 24 00:13:53.097989 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Nov 24 00:13:53.098001 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Nov 24 00:13:53.098012 kernel: TCP: Hash tables configured (established 32768 bind 32768) Nov 24 00:13:53.098033 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 24 00:13:53.098044 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 24 00:13:53.098070 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Nov 24 00:13:53.098082 kernel: NET: Registered PF_XDP protocol family Nov 24 00:13:53.098236 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Nov 24 00:13:53.098382 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Nov 24 00:13:53.098541 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Nov 24 00:13:53.098695 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Nov 24 00:13:53.098847 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Nov 24 00:13:53.098989 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Nov 24 00:13:53.099004 kernel: PCI: CLS 0 bytes, default 64 Nov 24 00:13:53.099016 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Nov 24 00:13:53.099038 kernel: Initialise system trusted keyrings Nov 24 00:13:53.099050 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Nov 24 00:13:53.099084 kernel: Key type asymmetric registered Nov 24 00:13:53.099095 kernel: Asymmetric key parser 'x509' registered Nov 24 00:13:53.099107 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Nov 24 00:13:53.099118 kernel: io scheduler mq-deadline registered Nov 24 00:13:53.099129 kernel: io scheduler kyber registered Nov 24 00:13:53.099140 kernel: io scheduler bfq registered Nov 24 00:13:53.099151 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Nov 24 00:13:53.099163 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Nov 24 00:13:53.099173 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Nov 24 00:13:53.099188 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Nov 24 00:13:53.099200 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Nov 24 00:13:53.099211 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Nov 24 00:13:53.099223 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Nov 24 00:13:53.099234 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Nov 24 00:13:53.099245 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Nov 24 00:13:53.099418 kernel: rtc_cmos 00:04: RTC can wake from S4 Nov 24 00:13:53.099436 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Nov 24 00:13:53.099603 kernel: rtc_cmos 00:04: registered as rtc0 Nov 24 00:13:53.099762 kernel: rtc_cmos 00:04: setting system clock to 2025-11-24T00:13:52 UTC (1763943232) Nov 24 00:13:53.099908 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Nov 24 00:13:53.099924 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Nov 24 00:13:53.099936 kernel: NET: Registered PF_INET6 protocol family Nov 24 00:13:53.099947 kernel: Segment Routing with IPv6 Nov 24 00:13:53.099958 kernel: In-situ OAM (IOAM) with IPv6 Nov 24 00:13:53.099969 kernel: NET: Registered PF_PACKET protocol family Nov 24 00:13:53.099980 kernel: Key type dns_resolver registered Nov 24 00:13:53.099996 kernel: IPI shorthand broadcast: enabled Nov 24 00:13:53.100007 kernel: sched_clock: Marking stable (3081003834, 280737386)->(3520950639, -159209419) Nov 24 00:13:53.100019 kernel: registered taskstats version 1 Nov 24 00:13:53.100040 kernel: Loading compiled-in X.509 certificates Nov 24 00:13:53.100052 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.58-flatcar: 5d380f93d180914be04be8068ab300f495c35900' Nov 24 00:13:53.100079 kernel: Demotion targets for Node 0: null Nov 24 00:13:53.100090 kernel: Key type .fscrypt registered Nov 24 00:13:53.100100 kernel: Key type fscrypt-provisioning registered Nov 24 00:13:53.100111 kernel: ima: No TPM chip found, activating TPM-bypass! Nov 24 00:13:53.100126 kernel: ima: Allocated hash algorithm: sha1 Nov 24 00:13:53.100138 kernel: ima: No architecture policies found Nov 24 00:13:53.100149 kernel: clk: Disabling unused clocks Nov 24 00:13:53.100160 kernel: Warning: unable to open an initial console. Nov 24 00:13:53.100171 kernel: Freeing unused kernel image (initmem) memory: 46188K Nov 24 00:13:53.100183 kernel: Write protecting the kernel read-only data: 40960k Nov 24 00:13:53.100194 kernel: Freeing unused kernel image (rodata/data gap) memory: 560K Nov 24 00:13:53.100205 kernel: Run /init as init process Nov 24 00:13:53.100215 kernel: with arguments: Nov 24 00:13:53.100229 kernel: /init Nov 24 00:13:53.100240 kernel: with environment: Nov 24 00:13:53.100251 kernel: HOME=/ Nov 24 00:13:53.100262 kernel: TERM=linux Nov 24 00:13:53.100274 systemd[1]: Successfully made /usr/ read-only. Nov 24 00:13:53.100289 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Nov 24 00:13:53.100318 systemd[1]: Detected virtualization kvm. Nov 24 00:13:53.100329 systemd[1]: Detected architecture x86-64. Nov 24 00:13:53.100341 systemd[1]: Running in initrd. Nov 24 00:13:53.100353 systemd[1]: No hostname configured, using default hostname. Nov 24 00:13:53.100365 systemd[1]: Hostname set to . Nov 24 00:13:53.100377 systemd[1]: Initializing machine ID from VM UUID. Nov 24 00:13:53.100388 systemd[1]: Queued start job for default target initrd.target. Nov 24 00:13:53.100400 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 24 00:13:53.100415 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 24 00:13:53.100428 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Nov 24 00:13:53.100440 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 24 00:13:53.100452 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Nov 24 00:13:53.100465 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Nov 24 00:13:53.100479 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Nov 24 00:13:53.100494 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Nov 24 00:13:53.100506 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 24 00:13:53.100518 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 24 00:13:53.100530 systemd[1]: Reached target paths.target - Path Units. Nov 24 00:13:53.100542 systemd[1]: Reached target slices.target - Slice Units. Nov 24 00:13:53.100553 systemd[1]: Reached target swap.target - Swaps. Nov 24 00:13:53.100565 systemd[1]: Reached target timers.target - Timer Units. Nov 24 00:13:53.100580 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Nov 24 00:13:53.100592 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 24 00:13:53.100607 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Nov 24 00:13:53.100619 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Nov 24 00:13:53.100631 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 24 00:13:53.100643 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 24 00:13:53.100655 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 24 00:13:53.100667 systemd[1]: Reached target sockets.target - Socket Units. Nov 24 00:13:53.100679 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Nov 24 00:13:53.100695 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 24 00:13:53.100707 systemd[1]: Finished network-cleanup.service - Network Cleanup. Nov 24 00:13:53.100719 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Nov 24 00:13:53.100731 systemd[1]: Starting systemd-fsck-usr.service... Nov 24 00:13:53.100743 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 24 00:13:53.100755 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 24 00:13:53.100767 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 24 00:13:53.100782 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Nov 24 00:13:53.100795 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 24 00:13:53.100842 systemd-journald[202]: Collecting audit messages is disabled. Nov 24 00:13:53.100875 systemd[1]: Finished systemd-fsck-usr.service. Nov 24 00:13:53.100888 systemd-journald[202]: Journal started Nov 24 00:13:53.100918 systemd-journald[202]: Runtime Journal (/run/log/journal/0e098e0e08db4405830d6975697df3a9) is 6M, max 48.3M, 42.2M free. Nov 24 00:13:53.110723 systemd-modules-load[203]: Inserted module 'overlay' Nov 24 00:13:53.114413 systemd[1]: Started systemd-journald.service - Journal Service. Nov 24 00:13:53.121199 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 24 00:13:53.205297 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 24 00:13:53.205329 kernel: Bridge firewalling registered Nov 24 00:13:53.147114 systemd-modules-load[203]: Inserted module 'br_netfilter' Nov 24 00:13:53.201899 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 24 00:13:53.206274 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 24 00:13:53.206980 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:13:53.260973 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 24 00:13:53.267233 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 24 00:13:53.272012 systemd-tmpfiles[216]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Nov 24 00:13:53.279350 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 24 00:13:53.284877 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 24 00:13:53.285967 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 24 00:13:53.290431 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 24 00:13:53.299394 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Nov 24 00:13:53.303292 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 24 00:13:53.307962 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 24 00:13:53.352130 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 24 00:13:53.361773 dracut-cmdline[238]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=1969a6ee0c0ec5507eb68849c160e94c58e52d2291c767873af68a1f52b30801 Nov 24 00:13:53.381280 systemd-resolved[239]: Positive Trust Anchors: Nov 24 00:13:53.381307 systemd-resolved[239]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 24 00:13:53.381346 systemd-resolved[239]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 24 00:13:53.384903 systemd-resolved[239]: Defaulting to hostname 'linux'. Nov 24 00:13:53.386648 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 24 00:13:53.400113 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 24 00:13:53.522099 kernel: SCSI subsystem initialized Nov 24 00:13:53.531091 kernel: Loading iSCSI transport class v2.0-870. Nov 24 00:13:53.542090 kernel: iscsi: registered transport (tcp) Nov 24 00:13:53.565095 kernel: iscsi: registered transport (qla4xxx) Nov 24 00:13:53.565148 kernel: QLogic iSCSI HBA Driver Nov 24 00:13:53.586564 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 24 00:13:53.608462 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 24 00:13:53.611230 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 24 00:13:53.669624 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Nov 24 00:13:53.680642 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Nov 24 00:13:53.744104 kernel: raid6: avx2x4 gen() 19385 MB/s Nov 24 00:13:53.761082 kernel: raid6: avx2x2 gen() 25319 MB/s Nov 24 00:13:53.779047 kernel: raid6: avx2x1 gen() 24578 MB/s Nov 24 00:13:53.779093 kernel: raid6: using algorithm avx2x2 gen() 25319 MB/s Nov 24 00:13:53.796877 kernel: raid6: .... xor() 16185 MB/s, rmw enabled Nov 24 00:13:53.796903 kernel: raid6: using avx2x2 recovery algorithm Nov 24 00:13:53.818093 kernel: xor: automatically using best checksumming function avx Nov 24 00:13:54.024114 kernel: Btrfs loaded, zoned=no, fsverity=no Nov 24 00:13:54.033277 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Nov 24 00:13:54.037089 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 24 00:13:54.069226 systemd-udevd[452]: Using default interface naming scheme 'v255'. Nov 24 00:13:54.075939 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 24 00:13:54.108688 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Nov 24 00:13:54.145672 dracut-pre-trigger[463]: rd.md=0: removing MD RAID activation Nov 24 00:13:54.177968 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Nov 24 00:13:54.195680 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 24 00:13:54.276687 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 24 00:13:54.281428 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Nov 24 00:13:54.353144 kernel: cryptd: max_cpu_qlen set to 1000 Nov 24 00:13:54.356110 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Nov 24 00:13:54.361353 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Nov 24 00:13:54.361591 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 24 00:13:54.361717 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:13:54.388182 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Nov 24 00:13:54.391541 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 24 00:13:54.402432 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Nov 24 00:13:54.402494 kernel: GPT:9289727 != 19775487 Nov 24 00:13:54.402508 kernel: GPT:Alternate GPT header not at the end of the disk. Nov 24 00:13:54.402522 kernel: GPT:9289727 != 19775487 Nov 24 00:13:54.402534 kernel: GPT: Use GNU Parted to correct GPT errors. Nov 24 00:13:54.402547 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 24 00:13:54.404910 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Nov 24 00:13:54.415096 kernel: libata version 3.00 loaded. Nov 24 00:13:54.422328 kernel: ahci 0000:00:1f.2: version 3.0 Nov 24 00:13:54.422570 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Nov 24 00:13:54.429197 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Nov 24 00:13:54.429379 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Nov 24 00:13:54.429527 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Nov 24 00:13:54.435079 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Nov 24 00:13:54.435104 kernel: AES CTR mode by8 optimization enabled Nov 24 00:13:54.436286 kernel: scsi host0: ahci Nov 24 00:13:54.453096 kernel: scsi host1: ahci Nov 24 00:13:54.476090 kernel: scsi host2: ahci Nov 24 00:13:54.476775 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Nov 24 00:13:54.538750 kernel: scsi host3: ahci Nov 24 00:13:54.539040 kernel: scsi host4: ahci Nov 24 00:13:54.539221 kernel: scsi host5: ahci Nov 24 00:13:54.539404 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 lpm-pol 1 Nov 24 00:13:54.539420 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 lpm-pol 1 Nov 24 00:13:54.539439 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 lpm-pol 1 Nov 24 00:13:54.539453 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 lpm-pol 1 Nov 24 00:13:54.539464 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 lpm-pol 1 Nov 24 00:13:54.539474 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 lpm-pol 1 Nov 24 00:13:54.542160 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:13:54.552095 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Nov 24 00:13:54.559723 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Nov 24 00:13:54.560370 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Nov 24 00:13:54.590868 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Nov 24 00:13:54.592418 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Nov 24 00:13:54.786089 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Nov 24 00:13:54.786185 kernel: ata3.00: LPM support broken, forcing max_power Nov 24 00:13:54.786201 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Nov 24 00:13:54.787282 kernel: ata3.00: applying bridge limits Nov 24 00:13:54.788073 kernel: ata5: SATA link down (SStatus 0 SControl 300) Nov 24 00:13:54.789107 kernel: ata6: SATA link down (SStatus 0 SControl 300) Nov 24 00:13:54.790124 kernel: ata3.00: LPM support broken, forcing max_power Nov 24 00:13:54.792262 kernel: ata3.00: configured for UDMA/100 Nov 24 00:13:54.793079 kernel: ata2: SATA link down (SStatus 0 SControl 300) Nov 24 00:13:54.794088 kernel: ata1: SATA link down (SStatus 0 SControl 300) Nov 24 00:13:54.804111 kernel: ata4: SATA link down (SStatus 0 SControl 300) Nov 24 00:13:54.804157 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Nov 24 00:13:54.870763 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Nov 24 00:13:54.871089 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Nov 24 00:13:54.883092 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Nov 24 00:13:55.046781 disk-uuid[614]: Primary Header is updated. Nov 24 00:13:55.046781 disk-uuid[614]: Secondary Entries is updated. Nov 24 00:13:55.046781 disk-uuid[614]: Secondary Header is updated. Nov 24 00:13:55.052634 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 24 00:13:55.055110 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 24 00:13:55.252578 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Nov 24 00:13:55.296201 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Nov 24 00:13:55.299823 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 24 00:13:55.331015 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 24 00:13:55.336192 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Nov 24 00:13:55.364756 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Nov 24 00:13:56.065177 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 24 00:13:56.065727 disk-uuid[617]: The operation has completed successfully. Nov 24 00:13:56.098549 systemd[1]: disk-uuid.service: Deactivated successfully. Nov 24 00:13:56.098716 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Nov 24 00:13:56.142619 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Nov 24 00:13:56.171052 sh[643]: Success Nov 24 00:13:56.190133 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Nov 24 00:13:56.190262 kernel: device-mapper: uevent: version 1.0.3 Nov 24 00:13:56.192606 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Nov 24 00:13:56.202152 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Nov 24 00:13:56.236624 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Nov 24 00:13:56.242523 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Nov 24 00:13:56.260145 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Nov 24 00:13:56.269102 kernel: BTRFS: device fsid c993ebd2-0e38-4cfc-8615-2c75294bea72 devid 1 transid 36 /dev/mapper/usr (253:0) scanned by mount (655) Nov 24 00:13:56.272756 kernel: BTRFS info (device dm-0): first mount of filesystem c993ebd2-0e38-4cfc-8615-2c75294bea72 Nov 24 00:13:56.272814 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Nov 24 00:13:56.280548 kernel: BTRFS info (device dm-0): disabling log replay at mount time Nov 24 00:13:56.280590 kernel: BTRFS info (device dm-0): enabling free space tree Nov 24 00:13:56.282170 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Nov 24 00:13:56.285722 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Nov 24 00:13:56.289480 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Nov 24 00:13:56.293513 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Nov 24 00:13:56.297913 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Nov 24 00:13:56.329117 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (688) Nov 24 00:13:56.332703 kernel: BTRFS info (device vda6): first mount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:13:56.332801 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 24 00:13:56.336990 kernel: BTRFS info (device vda6): turning on async discard Nov 24 00:13:56.337087 kernel: BTRFS info (device vda6): enabling free space tree Nov 24 00:13:56.344101 kernel: BTRFS info (device vda6): last unmount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:13:56.345680 systemd[1]: Finished ignition-setup.service - Ignition (setup). Nov 24 00:13:56.349024 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Nov 24 00:13:56.500575 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 24 00:13:56.526483 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 24 00:13:56.546602 ignition[733]: Ignition 2.22.0 Nov 24 00:13:56.546618 ignition[733]: Stage: fetch-offline Nov 24 00:13:56.546655 ignition[733]: no configs at "/usr/lib/ignition/base.d" Nov 24 00:13:56.546667 ignition[733]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 24 00:13:56.546763 ignition[733]: parsed url from cmdline: "" Nov 24 00:13:56.546768 ignition[733]: no config URL provided Nov 24 00:13:56.546774 ignition[733]: reading system config file "/usr/lib/ignition/user.ign" Nov 24 00:13:56.546785 ignition[733]: no config at "/usr/lib/ignition/user.ign" Nov 24 00:13:56.546814 ignition[733]: op(1): [started] loading QEMU firmware config module Nov 24 00:13:56.546821 ignition[733]: op(1): executing: "modprobe" "qemu_fw_cfg" Nov 24 00:13:56.575182 ignition[733]: op(1): [finished] loading QEMU firmware config module Nov 24 00:13:56.606871 systemd-networkd[829]: lo: Link UP Nov 24 00:13:56.606882 systemd-networkd[829]: lo: Gained carrier Nov 24 00:13:56.610127 systemd-networkd[829]: Enumeration completed Nov 24 00:13:56.610509 systemd-networkd[829]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 24 00:13:56.610514 systemd-networkd[829]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 24 00:13:56.610877 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 24 00:13:56.656580 systemd-networkd[829]: eth0: Link UP Nov 24 00:13:56.656822 systemd-networkd[829]: eth0: Gained carrier Nov 24 00:13:56.656838 systemd-networkd[829]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 24 00:13:56.661286 systemd[1]: Reached target network.target - Network. Nov 24 00:13:56.681208 systemd-networkd[829]: eth0: DHCPv4 address 10.0.0.25/16, gateway 10.0.0.1 acquired from 10.0.0.1 Nov 24 00:13:56.753608 ignition[733]: parsing config with SHA512: f4da17d43cc7f9360f7d8a0dc7f0ec37f37a54176cb97a600a2478b1fd890c15381cec2cafd2986d9ea5ca5413d659b5d79ed7f4663a7e5de6291105f46f9e9f Nov 24 00:13:56.757407 unknown[733]: fetched base config from "system" Nov 24 00:13:56.757418 unknown[733]: fetched user config from "qemu" Nov 24 00:13:56.757768 ignition[733]: fetch-offline: fetch-offline passed Nov 24 00:13:56.761069 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Nov 24 00:13:56.757829 ignition[733]: Ignition finished successfully Nov 24 00:13:56.793802 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Nov 24 00:13:56.796457 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Nov 24 00:13:56.940179 ignition[837]: Ignition 2.22.0 Nov 24 00:13:56.940197 ignition[837]: Stage: kargs Nov 24 00:13:56.940432 ignition[837]: no configs at "/usr/lib/ignition/base.d" Nov 24 00:13:56.940447 ignition[837]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 24 00:13:56.941398 ignition[837]: kargs: kargs passed Nov 24 00:13:56.941453 ignition[837]: Ignition finished successfully Nov 24 00:13:56.952687 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Nov 24 00:13:56.954859 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Nov 24 00:13:56.996540 ignition[845]: Ignition 2.22.0 Nov 24 00:13:56.996556 ignition[845]: Stage: disks Nov 24 00:13:56.996776 ignition[845]: no configs at "/usr/lib/ignition/base.d" Nov 24 00:13:56.996788 ignition[845]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 24 00:13:56.998127 ignition[845]: disks: disks passed Nov 24 00:13:56.998189 ignition[845]: Ignition finished successfully Nov 24 00:13:57.004592 systemd[1]: Finished ignition-disks.service - Ignition (disks). Nov 24 00:13:57.006687 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Nov 24 00:13:57.009744 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Nov 24 00:13:57.010693 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 24 00:13:57.019481 systemd[1]: Reached target sysinit.target - System Initialization. Nov 24 00:13:57.022708 systemd[1]: Reached target basic.target - Basic System. Nov 24 00:13:57.030110 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Nov 24 00:13:57.037728 systemd-resolved[239]: Detected conflict on linux IN A 10.0.0.25 Nov 24 00:13:57.037747 systemd-resolved[239]: Hostname conflict, changing published hostname from 'linux' to 'linux6'. Nov 24 00:13:57.067951 systemd-fsck[855]: ROOT: clean, 15/553520 files, 52789/553472 blocks Nov 24 00:13:57.212731 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Nov 24 00:13:57.216094 systemd[1]: Mounting sysroot.mount - /sysroot... Nov 24 00:13:57.349095 kernel: EXT4-fs (vda9): mounted filesystem 5d9d0447-100f-4769-adb5-76fdba966eb2 r/w with ordered data mode. Quota mode: none. Nov 24 00:13:57.349984 systemd[1]: Mounted sysroot.mount - /sysroot. Nov 24 00:13:57.351363 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Nov 24 00:13:57.356468 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 24 00:13:57.358358 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Nov 24 00:13:57.360328 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Nov 24 00:13:57.360382 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Nov 24 00:13:57.360409 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Nov 24 00:13:57.381994 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Nov 24 00:13:57.387038 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Nov 24 00:13:57.395090 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (863) Nov 24 00:13:57.399294 kernel: BTRFS info (device vda6): first mount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:13:57.399358 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 24 00:13:57.404407 kernel: BTRFS info (device vda6): turning on async discard Nov 24 00:13:57.404443 kernel: BTRFS info (device vda6): enabling free space tree Nov 24 00:13:57.405500 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 24 00:13:57.437748 initrd-setup-root[887]: cut: /sysroot/etc/passwd: No such file or directory Nov 24 00:13:57.444603 initrd-setup-root[894]: cut: /sysroot/etc/group: No such file or directory Nov 24 00:13:57.450730 initrd-setup-root[901]: cut: /sysroot/etc/shadow: No such file or directory Nov 24 00:13:57.457684 initrd-setup-root[908]: cut: /sysroot/etc/gshadow: No such file or directory Nov 24 00:13:57.590248 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Nov 24 00:13:57.592666 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Nov 24 00:13:57.595972 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Nov 24 00:13:57.614015 systemd[1]: sysroot-oem.mount: Deactivated successfully. Nov 24 00:13:57.617307 kernel: BTRFS info (device vda6): last unmount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:13:57.628636 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Nov 24 00:13:57.659709 ignition[978]: INFO : Ignition 2.22.0 Nov 24 00:13:57.659709 ignition[978]: INFO : Stage: mount Nov 24 00:13:57.662645 ignition[978]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 24 00:13:57.662645 ignition[978]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 24 00:13:57.662645 ignition[978]: INFO : mount: mount passed Nov 24 00:13:57.662645 ignition[978]: INFO : Ignition finished successfully Nov 24 00:13:57.664000 systemd[1]: Finished ignition-mount.service - Ignition (mount). Nov 24 00:13:57.667452 systemd[1]: Starting ignition-files.service - Ignition (files)... Nov 24 00:13:58.351949 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 24 00:13:58.373296 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (989) Nov 24 00:13:58.373347 kernel: BTRFS info (device vda6): first mount of filesystem 8f3e7759-f869-465c-a676-2cd550a2d4e4 Nov 24 00:13:58.373373 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 24 00:13:58.378809 kernel: BTRFS info (device vda6): turning on async discard Nov 24 00:13:58.378877 kernel: BTRFS info (device vda6): enabling free space tree Nov 24 00:13:58.380902 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 24 00:13:58.394215 systemd-networkd[829]: eth0: Gained IPv6LL Nov 24 00:13:58.422287 ignition[1006]: INFO : Ignition 2.22.0 Nov 24 00:13:58.422287 ignition[1006]: INFO : Stage: files Nov 24 00:13:58.425116 ignition[1006]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 24 00:13:58.425116 ignition[1006]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 24 00:13:58.430865 ignition[1006]: DEBUG : files: compiled without relabeling support, skipping Nov 24 00:13:58.433074 ignition[1006]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Nov 24 00:13:58.433074 ignition[1006]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Nov 24 00:13:58.440031 ignition[1006]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Nov 24 00:13:58.442730 ignition[1006]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Nov 24 00:13:58.445480 unknown[1006]: wrote ssh authorized keys file for user: core Nov 24 00:13:58.447318 ignition[1006]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Nov 24 00:13:58.451590 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 24 00:13:58.455155 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Nov 24 00:13:58.599984 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Nov 24 00:13:58.685455 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 24 00:13:58.685455 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Nov 24 00:13:58.692493 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 24 00:13:58.724404 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 24 00:13:58.724404 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 24 00:13:58.724404 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Nov 24 00:13:59.114453 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Nov 24 00:13:59.623527 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 24 00:13:59.623527 ignition[1006]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Nov 24 00:13:59.630778 ignition[1006]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 24 00:13:59.630778 ignition[1006]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 24 00:13:59.630778 ignition[1006]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Nov 24 00:13:59.630778 ignition[1006]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Nov 24 00:13:59.630778 ignition[1006]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Nov 24 00:13:59.630778 ignition[1006]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Nov 24 00:13:59.630778 ignition[1006]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Nov 24 00:13:59.630778 ignition[1006]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Nov 24 00:13:59.672489 ignition[1006]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Nov 24 00:13:59.676695 ignition[1006]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Nov 24 00:13:59.695960 ignition[1006]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Nov 24 00:13:59.695960 ignition[1006]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Nov 24 00:13:59.700705 ignition[1006]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Nov 24 00:13:59.700705 ignition[1006]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Nov 24 00:13:59.700705 ignition[1006]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Nov 24 00:13:59.700705 ignition[1006]: INFO : files: files passed Nov 24 00:13:59.700705 ignition[1006]: INFO : Ignition finished successfully Nov 24 00:13:59.704795 systemd[1]: Finished ignition-files.service - Ignition (files). Nov 24 00:13:59.713107 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Nov 24 00:13:59.714682 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Nov 24 00:13:59.750862 systemd[1]: ignition-quench.service: Deactivated successfully. Nov 24 00:13:59.750998 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Nov 24 00:13:59.756728 initrd-setup-root-after-ignition[1035]: grep: /sysroot/oem/oem-release: No such file or directory Nov 24 00:13:59.761141 initrd-setup-root-after-ignition[1037]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 24 00:13:59.761141 initrd-setup-root-after-ignition[1037]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Nov 24 00:13:59.766427 initrd-setup-root-after-ignition[1041]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 24 00:13:59.770948 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 24 00:13:59.771791 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Nov 24 00:13:59.778635 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Nov 24 00:13:59.838834 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Nov 24 00:13:59.838974 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Nov 24 00:13:59.843189 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Nov 24 00:13:59.847129 systemd[1]: Reached target initrd.target - Initrd Default Target. Nov 24 00:13:59.850831 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Nov 24 00:13:59.855550 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Nov 24 00:13:59.895473 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 24 00:13:59.898160 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Nov 24 00:13:59.928579 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Nov 24 00:13:59.929752 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 24 00:13:59.933548 systemd[1]: Stopped target timers.target - Timer Units. Nov 24 00:13:59.937758 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Nov 24 00:13:59.937936 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 24 00:13:59.944544 systemd[1]: Stopped target initrd.target - Initrd Default Target. Nov 24 00:13:59.945227 systemd[1]: Stopped target basic.target - Basic System. Nov 24 00:13:59.945833 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Nov 24 00:13:59.953681 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Nov 24 00:13:59.957595 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Nov 24 00:13:59.961614 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Nov 24 00:13:59.965649 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Nov 24 00:13:59.969734 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Nov 24 00:13:59.973094 systemd[1]: Stopped target sysinit.target - System Initialization. Nov 24 00:13:59.979700 systemd[1]: Stopped target local-fs.target - Local File Systems. Nov 24 00:13:59.980573 systemd[1]: Stopped target swap.target - Swaps. Nov 24 00:13:59.983750 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Nov 24 00:13:59.983995 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Nov 24 00:13:59.989130 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Nov 24 00:13:59.992591 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 24 00:13:59.996241 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Nov 24 00:13:59.996581 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 24 00:13:59.997540 systemd[1]: dracut-initqueue.service: Deactivated successfully. Nov 24 00:13:59.997696 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Nov 24 00:14:00.003777 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Nov 24 00:14:00.003955 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Nov 24 00:14:00.009707 systemd[1]: Stopped target paths.target - Path Units. Nov 24 00:14:00.013665 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Nov 24 00:14:00.020500 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 24 00:14:00.021568 systemd[1]: Stopped target slices.target - Slice Units. Nov 24 00:14:00.027663 systemd[1]: Stopped target sockets.target - Socket Units. Nov 24 00:14:00.028545 systemd[1]: iscsid.socket: Deactivated successfully. Nov 24 00:14:00.028671 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Nov 24 00:14:00.033954 systemd[1]: iscsiuio.socket: Deactivated successfully. Nov 24 00:14:00.034122 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 24 00:14:00.034776 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Nov 24 00:14:00.034957 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 24 00:14:00.039996 systemd[1]: ignition-files.service: Deactivated successfully. Nov 24 00:14:00.040157 systemd[1]: Stopped ignition-files.service - Ignition (files). Nov 24 00:14:00.044539 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Nov 24 00:14:00.046615 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Nov 24 00:14:00.046850 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Nov 24 00:14:00.081499 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Nov 24 00:14:00.084549 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Nov 24 00:14:00.086311 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Nov 24 00:14:00.090403 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Nov 24 00:14:00.090554 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Nov 24 00:14:00.100195 systemd[1]: initrd-cleanup.service: Deactivated successfully. Nov 24 00:14:00.100356 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Nov 24 00:14:00.120364 ignition[1061]: INFO : Ignition 2.22.0 Nov 24 00:14:00.120364 ignition[1061]: INFO : Stage: umount Nov 24 00:14:00.123333 ignition[1061]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 24 00:14:00.123333 ignition[1061]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 24 00:14:00.123333 ignition[1061]: INFO : umount: umount passed Nov 24 00:14:00.123333 ignition[1061]: INFO : Ignition finished successfully Nov 24 00:14:00.126628 systemd[1]: ignition-mount.service: Deactivated successfully. Nov 24 00:14:00.126808 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Nov 24 00:14:00.131721 systemd[1]: sysroot-boot.mount: Deactivated successfully. Nov 24 00:14:00.132345 systemd[1]: Stopped target network.target - Network. Nov 24 00:14:00.132805 systemd[1]: ignition-disks.service: Deactivated successfully. Nov 24 00:14:00.132872 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Nov 24 00:14:00.136668 systemd[1]: ignition-kargs.service: Deactivated successfully. Nov 24 00:14:00.136725 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Nov 24 00:14:00.140057 systemd[1]: ignition-setup.service: Deactivated successfully. Nov 24 00:14:00.140144 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Nov 24 00:14:00.143488 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Nov 24 00:14:00.143537 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Nov 24 00:14:00.146659 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Nov 24 00:14:00.150954 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Nov 24 00:14:00.160685 systemd[1]: sysroot-boot.service: Deactivated successfully. Nov 24 00:14:00.160858 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Nov 24 00:14:00.164802 systemd[1]: systemd-resolved.service: Deactivated successfully. Nov 24 00:14:00.164990 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Nov 24 00:14:00.170742 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Nov 24 00:14:00.171110 systemd[1]: systemd-networkd.service: Deactivated successfully. Nov 24 00:14:00.171319 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Nov 24 00:14:00.177421 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Nov 24 00:14:00.178801 systemd[1]: Stopped target network-pre.target - Preparation for Network. Nov 24 00:14:00.179875 systemd[1]: systemd-networkd.socket: Deactivated successfully. Nov 24 00:14:00.179954 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Nov 24 00:14:00.202532 systemd[1]: initrd-setup-root.service: Deactivated successfully. Nov 24 00:14:00.202637 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Nov 24 00:14:00.204941 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Nov 24 00:14:00.208691 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Nov 24 00:14:00.208755 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 24 00:14:00.212530 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 24 00:14:00.212592 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 24 00:14:00.219350 systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 24 00:14:00.219400 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Nov 24 00:14:00.223190 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Nov 24 00:14:00.223244 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 24 00:14:00.229145 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 24 00:14:00.234633 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Nov 24 00:14:00.234719 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Nov 24 00:14:00.256855 systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 24 00:14:00.260482 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 24 00:14:00.265933 systemd[1]: network-cleanup.service: Deactivated successfully. Nov 24 00:14:00.266118 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Nov 24 00:14:00.267959 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Nov 24 00:14:00.268055 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Nov 24 00:14:00.275136 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Nov 24 00:14:00.275199 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Nov 24 00:14:00.278990 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Nov 24 00:14:00.279120 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Nov 24 00:14:00.280955 systemd[1]: dracut-cmdline.service: Deactivated successfully. Nov 24 00:14:00.281034 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Nov 24 00:14:00.293326 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 24 00:14:00.293424 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 24 00:14:00.300961 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Nov 24 00:14:00.305044 systemd[1]: systemd-network-generator.service: Deactivated successfully. Nov 24 00:14:00.305145 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Nov 24 00:14:00.312118 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Nov 24 00:14:00.312237 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 24 00:14:00.318911 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 24 00:14:00.318996 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:14:00.326930 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Nov 24 00:14:00.327017 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Nov 24 00:14:00.327097 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Nov 24 00:14:00.341726 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Nov 24 00:14:00.341878 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Nov 24 00:14:00.346335 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Nov 24 00:14:00.351913 systemd[1]: Starting initrd-switch-root.service - Switch Root... Nov 24 00:14:00.376169 systemd[1]: Switching root. Nov 24 00:14:00.420110 systemd-journald[202]: Journal stopped Nov 24 00:14:02.226665 systemd-journald[202]: Received SIGTERM from PID 1 (systemd). Nov 24 00:14:02.226765 kernel: SELinux: policy capability network_peer_controls=1 Nov 24 00:14:02.226794 kernel: SELinux: policy capability open_perms=1 Nov 24 00:14:02.226819 kernel: SELinux: policy capability extended_socket_class=1 Nov 24 00:14:02.226842 kernel: SELinux: policy capability always_check_network=0 Nov 24 00:14:02.226858 kernel: SELinux: policy capability cgroup_seclabel=1 Nov 24 00:14:02.226874 kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 24 00:14:02.226896 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Nov 24 00:14:02.226912 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Nov 24 00:14:02.226927 kernel: SELinux: policy capability userspace_initial_context=0 Nov 24 00:14:02.226943 kernel: audit: type=1403 audit(1763943241.166:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Nov 24 00:14:02.226962 systemd[1]: Successfully loaded SELinux policy in 76.649ms. Nov 24 00:14:02.226995 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 11.460ms. Nov 24 00:14:02.227014 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Nov 24 00:14:02.227032 systemd[1]: Detected virtualization kvm. Nov 24 00:14:02.227050 systemd[1]: Detected architecture x86-64. Nov 24 00:14:02.227093 systemd[1]: Detected first boot. Nov 24 00:14:02.227110 systemd[1]: Initializing machine ID from VM UUID. Nov 24 00:14:02.227127 zram_generator::config[1111]: No configuration found. Nov 24 00:14:02.227144 kernel: Guest personality initialized and is inactive Nov 24 00:14:02.227160 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Nov 24 00:14:02.227176 kernel: Initialized host personality Nov 24 00:14:02.227192 kernel: NET: Registered PF_VSOCK protocol family Nov 24 00:14:02.227208 systemd[1]: Populated /etc with preset unit settings. Nov 24 00:14:02.227231 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Nov 24 00:14:02.227253 systemd[1]: initrd-switch-root.service: Deactivated successfully. Nov 24 00:14:02.227269 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Nov 24 00:14:02.227285 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Nov 24 00:14:02.227302 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Nov 24 00:14:02.227319 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Nov 24 00:14:02.227343 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Nov 24 00:14:02.227360 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Nov 24 00:14:02.227377 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Nov 24 00:14:02.227397 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Nov 24 00:14:02.227415 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Nov 24 00:14:02.227431 systemd[1]: Created slice user.slice - User and Session Slice. Nov 24 00:14:02.227448 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 24 00:14:02.227465 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 24 00:14:02.227481 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Nov 24 00:14:02.227498 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Nov 24 00:14:02.227515 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Nov 24 00:14:02.227542 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 24 00:14:02.227562 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Nov 24 00:14:02.227579 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 24 00:14:02.227596 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 24 00:14:02.227612 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Nov 24 00:14:02.227632 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Nov 24 00:14:02.227648 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Nov 24 00:14:02.227665 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Nov 24 00:14:02.227686 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 24 00:14:02.227702 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 24 00:14:02.227719 systemd[1]: Reached target slices.target - Slice Units. Nov 24 00:14:02.227736 systemd[1]: Reached target swap.target - Swaps. Nov 24 00:14:02.227752 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Nov 24 00:14:02.227769 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Nov 24 00:14:02.227796 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Nov 24 00:14:02.227814 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 24 00:14:02.227830 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 24 00:14:02.227846 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 24 00:14:02.227867 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Nov 24 00:14:02.227884 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Nov 24 00:14:02.227900 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Nov 24 00:14:02.227923 systemd[1]: Mounting media.mount - External Media Directory... Nov 24 00:14:02.227940 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:14:02.227955 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Nov 24 00:14:02.227970 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Nov 24 00:14:02.227985 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Nov 24 00:14:02.228007 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Nov 24 00:14:02.228023 systemd[1]: Reached target machines.target - Containers. Nov 24 00:14:02.228041 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Nov 24 00:14:02.228078 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 24 00:14:02.228097 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 24 00:14:02.228114 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Nov 24 00:14:02.228130 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 24 00:14:02.228147 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 24 00:14:02.228163 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 24 00:14:02.228185 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Nov 24 00:14:02.228202 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 24 00:14:02.228220 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Nov 24 00:14:02.228236 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Nov 24 00:14:02.228253 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Nov 24 00:14:02.228269 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Nov 24 00:14:02.228286 systemd[1]: Stopped systemd-fsck-usr.service. Nov 24 00:14:02.228304 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 24 00:14:02.228325 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 24 00:14:02.228341 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 24 00:14:02.228357 kernel: fuse: init (API version 7.41) Nov 24 00:14:02.228373 kernel: loop: module loaded Nov 24 00:14:02.228392 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 24 00:14:02.228408 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Nov 24 00:14:02.228425 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Nov 24 00:14:02.228441 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 24 00:14:02.228459 systemd[1]: verity-setup.service: Deactivated successfully. Nov 24 00:14:02.228479 systemd[1]: Stopped verity-setup.service. Nov 24 00:14:02.228497 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:14:02.228514 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Nov 24 00:14:02.228531 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Nov 24 00:14:02.228548 systemd[1]: Mounted media.mount - External Media Directory. Nov 24 00:14:02.228611 systemd-journald[1171]: Collecting audit messages is disabled. Nov 24 00:14:02.228643 kernel: ACPI: bus type drm_connector registered Nov 24 00:14:02.228660 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Nov 24 00:14:02.228676 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Nov 24 00:14:02.228697 systemd-journald[1171]: Journal started Nov 24 00:14:02.228727 systemd-journald[1171]: Runtime Journal (/run/log/journal/0e098e0e08db4405830d6975697df3a9) is 6M, max 48.3M, 42.2M free. Nov 24 00:14:01.842799 systemd[1]: Queued start job for default target multi-user.target. Nov 24 00:14:01.865319 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Nov 24 00:14:01.866037 systemd[1]: systemd-journald.service: Deactivated successfully. Nov 24 00:14:02.234113 systemd[1]: Started systemd-journald.service - Journal Service. Nov 24 00:14:02.235219 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Nov 24 00:14:02.237510 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 24 00:14:02.241500 systemd[1]: modprobe@configfs.service: Deactivated successfully. Nov 24 00:14:02.241806 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Nov 24 00:14:02.244318 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 24 00:14:02.244605 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 24 00:14:02.246956 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 24 00:14:02.247249 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 24 00:14:02.250108 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 24 00:14:02.250387 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 24 00:14:02.252882 systemd[1]: modprobe@fuse.service: Deactivated successfully. Nov 24 00:14:02.253196 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Nov 24 00:14:02.255628 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 24 00:14:02.255912 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 24 00:14:02.258434 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 24 00:14:02.260812 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 24 00:14:02.263516 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Nov 24 00:14:02.266299 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Nov 24 00:14:02.287112 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 24 00:14:02.292761 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Nov 24 00:14:02.299234 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Nov 24 00:14:02.301483 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Nov 24 00:14:02.301546 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 24 00:14:02.304835 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Nov 24 00:14:02.311311 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Nov 24 00:14:02.314394 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 24 00:14:02.317654 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Nov 24 00:14:02.319985 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Nov 24 00:14:02.321992 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 24 00:14:02.332608 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Nov 24 00:14:02.337006 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 24 00:14:02.340192 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 24 00:14:02.345209 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Nov 24 00:14:02.347513 systemd-journald[1171]: Time spent on flushing to /var/log/journal/0e098e0e08db4405830d6975697df3a9 is 28.968ms for 982 entries. Nov 24 00:14:02.347513 systemd-journald[1171]: System Journal (/var/log/journal/0e098e0e08db4405830d6975697df3a9) is 8M, max 195.6M, 187.6M free. Nov 24 00:14:02.389104 systemd-journald[1171]: Received client request to flush runtime journal. Nov 24 00:14:02.350687 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Nov 24 00:14:02.365847 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 24 00:14:02.368616 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Nov 24 00:14:02.371011 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Nov 24 00:14:02.373816 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Nov 24 00:14:02.380549 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Nov 24 00:14:02.388223 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Nov 24 00:14:02.392262 systemd[1]: Starting systemd-sysusers.service - Create System Users... Nov 24 00:14:02.394670 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Nov 24 00:14:02.437091 kernel: loop0: detected capacity change from 0 to 224512 Nov 24 00:14:02.438894 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Nov 24 00:14:02.449683 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Nov 24 00:14:02.457599 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 24 00:14:02.460093 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Nov 24 00:14:02.475853 systemd[1]: Finished systemd-sysusers.service - Create System Users. Nov 24 00:14:02.481222 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 24 00:14:02.484170 kernel: loop1: detected capacity change from 0 to 110984 Nov 24 00:14:02.511626 systemd-tmpfiles[1243]: ACLs are not supported, ignoring. Nov 24 00:14:02.511651 systemd-tmpfiles[1243]: ACLs are not supported, ignoring. Nov 24 00:14:02.516136 kernel: loop2: detected capacity change from 0 to 128560 Nov 24 00:14:02.518526 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 24 00:14:02.636098 kernel: loop3: detected capacity change from 0 to 224512 Nov 24 00:14:02.653087 kernel: loop4: detected capacity change from 0 to 110984 Nov 24 00:14:02.664089 kernel: loop5: detected capacity change from 0 to 128560 Nov 24 00:14:02.670075 (sd-merge)[1248]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Nov 24 00:14:02.670999 (sd-merge)[1248]: Merged extensions into '/usr'. Nov 24 00:14:02.682809 systemd[1]: Reload requested from client PID 1225 ('systemd-sysext') (unit systemd-sysext.service)... Nov 24 00:14:02.682832 systemd[1]: Reloading... Nov 24 00:14:02.880107 zram_generator::config[1274]: No configuration found. Nov 24 00:14:03.070575 ldconfig[1220]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Nov 24 00:14:03.196127 systemd[1]: Reloading finished in 512 ms. Nov 24 00:14:03.214564 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Nov 24 00:14:03.217257 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Nov 24 00:14:03.244557 systemd[1]: Starting ensure-sysext.service... Nov 24 00:14:03.249636 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 24 00:14:03.280946 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Nov 24 00:14:03.286192 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 24 00:14:03.292405 systemd[1]: Reload requested from client PID 1312 ('systemctl') (unit ensure-sysext.service)... Nov 24 00:14:03.292525 systemd[1]: Reloading... Nov 24 00:14:03.294904 systemd-tmpfiles[1313]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Nov 24 00:14:03.294950 systemd-tmpfiles[1313]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Nov 24 00:14:03.295306 systemd-tmpfiles[1313]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Nov 24 00:14:03.295579 systemd-tmpfiles[1313]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Nov 24 00:14:03.296591 systemd-tmpfiles[1313]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Nov 24 00:14:03.296881 systemd-tmpfiles[1313]: ACLs are not supported, ignoring. Nov 24 00:14:03.296965 systemd-tmpfiles[1313]: ACLs are not supported, ignoring. Nov 24 00:14:03.301560 systemd-tmpfiles[1313]: Detected autofs mount point /boot during canonicalization of boot. Nov 24 00:14:03.301572 systemd-tmpfiles[1313]: Skipping /boot Nov 24 00:14:03.312778 systemd-tmpfiles[1313]: Detected autofs mount point /boot during canonicalization of boot. Nov 24 00:14:03.312792 systemd-tmpfiles[1313]: Skipping /boot Nov 24 00:14:03.380360 systemd-udevd[1316]: Using default interface naming scheme 'v255'. Nov 24 00:14:03.384118 zram_generator::config[1341]: No configuration found. Nov 24 00:14:03.641164 kernel: mousedev: PS/2 mouse device common for all mice Nov 24 00:14:03.666097 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Nov 24 00:14:03.671249 systemd[1]: Reloading finished in 378 ms. Nov 24 00:14:03.672135 kernel: ACPI: button: Power Button [PWRF] Nov 24 00:14:03.677129 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Nov 24 00:14:03.677512 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Nov 24 00:14:03.687374 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 24 00:14:03.690348 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 24 00:14:03.758701 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Nov 24 00:14:03.766467 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Nov 24 00:14:03.785735 systemd[1]: Finished ensure-sysext.service. Nov 24 00:14:03.794789 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:14:03.798219 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 24 00:14:03.805177 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Nov 24 00:14:03.807281 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 24 00:14:03.811308 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 24 00:14:03.834133 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 24 00:14:03.842259 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 24 00:14:03.845514 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 24 00:14:03.847628 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 24 00:14:03.849324 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Nov 24 00:14:03.851529 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 24 00:14:03.853337 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Nov 24 00:14:03.874376 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 24 00:14:03.881337 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 24 00:14:03.889393 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Nov 24 00:14:03.920613 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Nov 24 00:14:03.927251 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 24 00:14:03.930160 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 24 00:14:03.933101 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Nov 24 00:14:03.936163 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 24 00:14:03.952431 augenrules[1464]: No rules Nov 24 00:14:03.979870 kernel: kvm_amd: TSC scaling supported Nov 24 00:14:03.979944 kernel: kvm_amd: Nested Virtualization enabled Nov 24 00:14:03.979961 kernel: kvm_amd: Nested Paging enabled Nov 24 00:14:03.981718 kernel: kvm_amd: LBR virtualization supported Nov 24 00:14:03.981869 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Nov 24 00:14:03.986105 kernel: kvm_amd: Virtual GIF supported Nov 24 00:14:03.982439 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 24 00:14:03.986541 systemd[1]: audit-rules.service: Deactivated successfully. Nov 24 00:14:03.987240 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 24 00:14:03.990267 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 24 00:14:03.990896 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 24 00:14:03.996959 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 24 00:14:03.997503 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 24 00:14:04.000693 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 24 00:14:04.001051 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 24 00:14:04.004171 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Nov 24 00:14:04.005374 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Nov 24 00:14:04.016157 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 24 00:14:04.016353 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 24 00:14:04.018173 systemd[1]: Starting systemd-update-done.service - Update is Completed... Nov 24 00:14:04.023254 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Nov 24 00:14:04.023932 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Nov 24 00:14:04.024521 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Nov 24 00:14:04.040117 kernel: EDAC MC: Ver: 3.0.0 Nov 24 00:14:04.069190 systemd[1]: Finished systemd-update-done.service - Update is Completed. Nov 24 00:14:04.104978 systemd[1]: Started systemd-userdbd.service - User Database Manager. Nov 24 00:14:04.149435 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 24 00:14:04.180760 systemd-networkd[1448]: lo: Link UP Nov 24 00:14:04.180770 systemd-networkd[1448]: lo: Gained carrier Nov 24 00:14:04.182527 systemd-networkd[1448]: Enumeration completed Nov 24 00:14:04.182698 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 24 00:14:04.183902 systemd-networkd[1448]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 24 00:14:04.183914 systemd-networkd[1448]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 24 00:14:04.184591 systemd-networkd[1448]: eth0: Link UP Nov 24 00:14:04.186444 systemd-networkd[1448]: eth0: Gained carrier Nov 24 00:14:04.186481 systemd-networkd[1448]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 24 00:14:04.189766 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Nov 24 00:14:04.192931 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Nov 24 00:14:04.194873 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Nov 24 00:14:04.197399 systemd[1]: Reached target time-set.target - System Time Set. Nov 24 00:14:04.203244 systemd-resolved[1450]: Positive Trust Anchors: Nov 24 00:14:04.203294 systemd-resolved[1450]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 24 00:14:04.203331 systemd-resolved[1450]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 24 00:14:04.207131 systemd-networkd[1448]: eth0: DHCPv4 address 10.0.0.25/16, gateway 10.0.0.1 acquired from 10.0.0.1 Nov 24 00:14:04.208665 systemd-resolved[1450]: Defaulting to hostname 'linux'. Nov 24 00:14:04.208766 systemd-timesyncd[1454]: Network configuration changed, trying to establish connection. Nov 24 00:14:05.142078 systemd-timesyncd[1454]: Contacted time server 10.0.0.1:123 (10.0.0.1). Nov 24 00:14:05.142214 systemd-timesyncd[1454]: Initial clock synchronization to Mon 2025-11-24 00:14:05.141788 UTC. Nov 24 00:14:05.143100 systemd-resolved[1450]: Clock change detected. Flushing caches. Nov 24 00:14:05.143156 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 24 00:14:05.145257 systemd[1]: Reached target network.target - Network. Nov 24 00:14:05.146832 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 24 00:14:05.148982 systemd[1]: Reached target sysinit.target - System Initialization. Nov 24 00:14:05.150837 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Nov 24 00:14:05.152847 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Nov 24 00:14:05.154925 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Nov 24 00:14:05.157255 systemd[1]: Started logrotate.timer - Daily rotation of log files. Nov 24 00:14:05.159170 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Nov 24 00:14:05.161283 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Nov 24 00:14:05.163428 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Nov 24 00:14:05.163466 systemd[1]: Reached target paths.target - Path Units. Nov 24 00:14:05.164978 systemd[1]: Reached target timers.target - Timer Units. Nov 24 00:14:05.167806 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Nov 24 00:14:05.171475 systemd[1]: Starting docker.socket - Docker Socket for the API... Nov 24 00:14:05.175965 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Nov 24 00:14:05.178297 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Nov 24 00:14:05.180372 systemd[1]: Reached target ssh-access.target - SSH Access Available. Nov 24 00:14:05.186530 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Nov 24 00:14:05.188582 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Nov 24 00:14:05.191383 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Nov 24 00:14:05.193672 systemd[1]: Listening on docker.socket - Docker Socket for the API. Nov 24 00:14:05.196935 systemd[1]: Reached target sockets.target - Socket Units. Nov 24 00:14:05.198567 systemd[1]: Reached target basic.target - Basic System. Nov 24 00:14:05.200201 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Nov 24 00:14:05.200251 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Nov 24 00:14:05.201669 systemd[1]: Starting containerd.service - containerd container runtime... Nov 24 00:14:05.204523 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Nov 24 00:14:05.207096 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Nov 24 00:14:05.209558 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Nov 24 00:14:05.218840 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Nov 24 00:14:05.220914 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Nov 24 00:14:05.222258 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Nov 24 00:14:05.227919 jq[1504]: false Nov 24 00:14:05.226103 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Nov 24 00:14:05.228137 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Nov 24 00:14:05.231830 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Nov 24 00:14:05.235924 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Nov 24 00:14:05.242866 google_oslogin_nss_cache[1506]: oslogin_cache_refresh[1506]: Refreshing passwd entry cache Nov 24 00:14:05.242807 oslogin_cache_refresh[1506]: Refreshing passwd entry cache Nov 24 00:14:05.246803 systemd[1]: Starting systemd-logind.service - User Login Management... Nov 24 00:14:05.251162 extend-filesystems[1505]: Found /dev/vda6 Nov 24 00:14:05.250722 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Nov 24 00:14:05.253607 extend-filesystems[1505]: Found /dev/vda9 Nov 24 00:14:05.252954 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Nov 24 00:14:05.253699 oslogin_cache_refresh[1506]: Failure getting users, quitting Nov 24 00:14:05.255355 extend-filesystems[1505]: Checking size of /dev/vda9 Nov 24 00:14:05.256058 google_oslogin_nss_cache[1506]: oslogin_cache_refresh[1506]: Failure getting users, quitting Nov 24 00:14:05.256058 google_oslogin_nss_cache[1506]: oslogin_cache_refresh[1506]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Nov 24 00:14:05.256058 google_oslogin_nss_cache[1506]: oslogin_cache_refresh[1506]: Refreshing group entry cache Nov 24 00:14:05.254445 systemd[1]: Starting update-engine.service - Update Engine... Nov 24 00:14:05.253722 oslogin_cache_refresh[1506]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Nov 24 00:14:05.253789 oslogin_cache_refresh[1506]: Refreshing group entry cache Nov 24 00:14:05.260746 google_oslogin_nss_cache[1506]: oslogin_cache_refresh[1506]: Failure getting groups, quitting Nov 24 00:14:05.260746 google_oslogin_nss_cache[1506]: oslogin_cache_refresh[1506]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Nov 24 00:14:05.260724 oslogin_cache_refresh[1506]: Failure getting groups, quitting Nov 24 00:14:05.260737 oslogin_cache_refresh[1506]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Nov 24 00:14:05.260973 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Nov 24 00:14:05.267108 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Nov 24 00:14:05.270319 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Nov 24 00:14:05.270585 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Nov 24 00:14:05.270938 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Nov 24 00:14:05.271815 jq[1524]: true Nov 24 00:14:05.271183 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Nov 24 00:14:05.273596 systemd[1]: motdgen.service: Deactivated successfully. Nov 24 00:14:05.273873 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Nov 24 00:14:05.278039 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Nov 24 00:14:05.280979 extend-filesystems[1505]: Resized partition /dev/vda9 Nov 24 00:14:05.282901 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Nov 24 00:14:05.294936 extend-filesystems[1534]: resize2fs 1.47.3 (8-Jul-2025) Nov 24 00:14:05.300120 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Nov 24 00:14:05.300351 jq[1536]: true Nov 24 00:14:05.305343 update_engine[1522]: I20251124 00:14:05.302731 1522 main.cc:92] Flatcar Update Engine starting Nov 24 00:14:05.313430 tar[1529]: linux-amd64/LICENSE Nov 24 00:14:05.314756 tar[1529]: linux-amd64/helm Nov 24 00:14:05.317079 (ntainerd)[1537]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Nov 24 00:14:05.336672 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Nov 24 00:14:05.358835 systemd-logind[1519]: Watching system buttons on /dev/input/event2 (Power Button) Nov 24 00:14:05.360436 extend-filesystems[1534]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Nov 24 00:14:05.360436 extend-filesystems[1534]: old_desc_blocks = 1, new_desc_blocks = 1 Nov 24 00:14:05.360436 extend-filesystems[1534]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Nov 24 00:14:05.358874 systemd-logind[1519]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Nov 24 00:14:05.374846 extend-filesystems[1505]: Resized filesystem in /dev/vda9 Nov 24 00:14:05.359369 systemd-logind[1519]: New seat seat0. Nov 24 00:14:05.361258 systemd[1]: Started systemd-logind.service - User Login Management. Nov 24 00:14:05.364664 systemd[1]: extend-filesystems.service: Deactivated successfully. Nov 24 00:14:05.365037 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Nov 24 00:14:05.388171 dbus-daemon[1502]: [system] SELinux support is enabled Nov 24 00:14:05.388396 systemd[1]: Started dbus.service - D-Bus System Message Bus. Nov 24 00:14:05.394780 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Nov 24 00:14:05.394822 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Nov 24 00:14:05.397250 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Nov 24 00:14:05.397274 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Nov 24 00:14:05.408142 dbus-daemon[1502]: [system] Successfully activated service 'org.freedesktop.systemd1' Nov 24 00:14:05.411145 systemd[1]: Started update-engine.service - Update Engine. Nov 24 00:14:05.415805 update_engine[1522]: I20251124 00:14:05.412364 1522 update_check_scheduler.cc:74] Next update check in 3m58s Nov 24 00:14:05.417036 systemd[1]: Started locksmithd.service - Cluster reboot manager. Nov 24 00:14:05.431082 bash[1564]: Updated "/home/core/.ssh/authorized_keys" Nov 24 00:14:05.432977 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Nov 24 00:14:05.433732 sshd_keygen[1535]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Nov 24 00:14:05.449555 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Nov 24 00:14:05.480228 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Nov 24 00:14:05.486814 systemd[1]: Starting issuegen.service - Generate /run/issue... Nov 24 00:14:05.533939 locksmithd[1568]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Nov 24 00:14:05.574795 systemd[1]: issuegen.service: Deactivated successfully. Nov 24 00:14:05.575111 systemd[1]: Finished issuegen.service - Generate /run/issue. Nov 24 00:14:05.581030 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Nov 24 00:14:05.635019 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Nov 24 00:14:05.638991 systemd[1]: Started getty@tty1.service - Getty on tty1. Nov 24 00:14:05.644026 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Nov 24 00:14:05.646380 systemd[1]: Reached target getty.target - Login Prompts. Nov 24 00:14:05.705748 containerd[1537]: time="2025-11-24T00:14:05Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Nov 24 00:14:05.706732 containerd[1537]: time="2025-11-24T00:14:05.706691245Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Nov 24 00:14:05.724650 containerd[1537]: time="2025-11-24T00:14:05.724569513Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="20.699µs" Nov 24 00:14:05.724735 containerd[1537]: time="2025-11-24T00:14:05.724648671Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Nov 24 00:14:05.724735 containerd[1537]: time="2025-11-24T00:14:05.724691622Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Nov 24 00:14:05.725035 containerd[1537]: time="2025-11-24T00:14:05.725000331Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Nov 24 00:14:05.725035 containerd[1537]: time="2025-11-24T00:14:05.725030177Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Nov 24 00:14:05.725103 containerd[1537]: time="2025-11-24T00:14:05.725074740Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Nov 24 00:14:05.725218 containerd[1537]: time="2025-11-24T00:14:05.725183715Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Nov 24 00:14:05.725218 containerd[1537]: time="2025-11-24T00:14:05.725205906Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Nov 24 00:14:05.725721 containerd[1537]: time="2025-11-24T00:14:05.725683352Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Nov 24 00:14:05.725721 containerd[1537]: time="2025-11-24T00:14:05.725710362Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Nov 24 00:14:05.725765 containerd[1537]: time="2025-11-24T00:14:05.725727324Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Nov 24 00:14:05.725765 containerd[1537]: time="2025-11-24T00:14:05.725739998Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Nov 24 00:14:05.725930 containerd[1537]: time="2025-11-24T00:14:05.725898305Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Nov 24 00:14:05.726332 containerd[1537]: time="2025-11-24T00:14:05.726298175Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Nov 24 00:14:05.726375 containerd[1537]: time="2025-11-24T00:14:05.726351525Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Nov 24 00:14:05.726375 containerd[1537]: time="2025-11-24T00:14:05.726370230Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Nov 24 00:14:05.726483 containerd[1537]: time="2025-11-24T00:14:05.726450861Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Nov 24 00:14:05.726952 containerd[1537]: time="2025-11-24T00:14:05.726918648Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Nov 24 00:14:05.727033 containerd[1537]: time="2025-11-24T00:14:05.727013336Z" level=info msg="metadata content store policy set" policy=shared Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.733934826Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734227465Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734258303Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734287467Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734309579Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734326841Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734341328Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734369601Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734389459Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734405829Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734420497Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Nov 24 00:14:05.734650 containerd[1537]: time="2025-11-24T00:14:05.734446836Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Nov 24 00:14:05.734913 containerd[1537]: time="2025-11-24T00:14:05.734893684Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Nov 24 00:14:05.735010 containerd[1537]: time="2025-11-24T00:14:05.734994183Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Nov 24 00:14:05.735072 containerd[1537]: time="2025-11-24T00:14:05.735058974Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Nov 24 00:14:05.735124 containerd[1537]: time="2025-11-24T00:14:05.735109619Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Nov 24 00:14:05.735177 containerd[1537]: time="2025-11-24T00:14:05.735162799Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Nov 24 00:14:05.735234 containerd[1537]: time="2025-11-24T00:14:05.735221769Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Nov 24 00:14:05.735311 containerd[1537]: time="2025-11-24T00:14:05.735296209Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Nov 24 00:14:05.735383 containerd[1537]: time="2025-11-24T00:14:05.735363886Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Nov 24 00:14:05.735459 containerd[1537]: time="2025-11-24T00:14:05.735438626Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Nov 24 00:14:05.735530 containerd[1537]: time="2025-11-24T00:14:05.735511713Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Nov 24 00:14:05.735603 containerd[1537]: time="2025-11-24T00:14:05.735586093Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Nov 24 00:14:05.735806 containerd[1537]: time="2025-11-24T00:14:05.735780036Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Nov 24 00:14:05.735944 containerd[1537]: time="2025-11-24T00:14:05.735910972Z" level=info msg="Start snapshots syncer" Nov 24 00:14:05.736006 containerd[1537]: time="2025-11-24T00:14:05.735974832Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Nov 24 00:14:05.736899 containerd[1537]: time="2025-11-24T00:14:05.736807163Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Nov 24 00:14:05.737110 containerd[1537]: time="2025-11-24T00:14:05.736939451Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Nov 24 00:14:05.737110 containerd[1537]: time="2025-11-24T00:14:05.737061069Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Nov 24 00:14:05.737312 containerd[1537]: time="2025-11-24T00:14:05.737260232Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Nov 24 00:14:05.737426 containerd[1537]: time="2025-11-24T00:14:05.737340112Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Nov 24 00:14:05.737465 containerd[1537]: time="2025-11-24T00:14:05.737438056Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Nov 24 00:14:05.737500 containerd[1537]: time="2025-11-24T00:14:05.737474965Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Nov 24 00:14:05.737549 containerd[1537]: time="2025-11-24T00:14:05.737513798Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Nov 24 00:14:05.737593 containerd[1537]: time="2025-11-24T00:14:05.737566366Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Nov 24 00:14:05.737656 containerd[1537]: time="2025-11-24T00:14:05.737602524Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Nov 24 00:14:05.737720 containerd[1537]: time="2025-11-24T00:14:05.737689287Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Nov 24 00:14:05.737742 containerd[1537]: time="2025-11-24T00:14:05.737726627Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Nov 24 00:14:05.737761 containerd[1537]: time="2025-11-24T00:14:05.737752475Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Nov 24 00:14:05.737878 containerd[1537]: time="2025-11-24T00:14:05.737859176Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Nov 24 00:14:05.737905 containerd[1537]: time="2025-11-24T00:14:05.737885535Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Nov 24 00:14:05.737940 containerd[1537]: time="2025-11-24T00:14:05.737912145Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Nov 24 00:14:05.737995 containerd[1537]: time="2025-11-24T00:14:05.737950637Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Nov 24 00:14:05.737995 containerd[1537]: time="2025-11-24T00:14:05.737978429Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Nov 24 00:14:05.738075 containerd[1537]: time="2025-11-24T00:14:05.738027080Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Nov 24 00:14:05.738118 containerd[1537]: time="2025-11-24T00:14:05.738099176Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Nov 24 00:14:05.738161 containerd[1537]: time="2025-11-24T00:14:05.738141776Z" level=info msg="runtime interface created" Nov 24 00:14:05.738181 containerd[1537]: time="2025-11-24T00:14:05.738160911Z" level=info msg="created NRI interface" Nov 24 00:14:05.738232 containerd[1537]: time="2025-11-24T00:14:05.738182923Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Nov 24 00:14:05.738232 containerd[1537]: time="2025-11-24T00:14:05.738213720Z" level=info msg="Connect containerd service" Nov 24 00:14:05.738289 containerd[1537]: time="2025-11-24T00:14:05.738259045Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Nov 24 00:14:05.740076 containerd[1537]: time="2025-11-24T00:14:05.740041348Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 24 00:14:05.974782 tar[1529]: linux-amd64/README.md Nov 24 00:14:06.005653 containerd[1537]: time="2025-11-24T00:14:06.004478462Z" level=info msg="Start subscribing containerd event" Nov 24 00:14:06.005653 containerd[1537]: time="2025-11-24T00:14:06.004842014Z" level=info msg="Start recovering state" Nov 24 00:14:06.005954 containerd[1537]: time="2025-11-24T00:14:06.005896581Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Nov 24 00:14:06.006036 containerd[1537]: time="2025-11-24T00:14:06.006002189Z" level=info msg="Start event monitor" Nov 24 00:14:06.006290 containerd[1537]: time="2025-11-24T00:14:06.006227963Z" level=info msg="Start cni network conf syncer for default" Nov 24 00:14:06.006290 containerd[1537]: time="2025-11-24T00:14:06.006258771Z" level=info msg="Start streaming server" Nov 24 00:14:06.006290 containerd[1537]: time="2025-11-24T00:14:06.006297694Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Nov 24 00:14:06.006475 containerd[1537]: time="2025-11-24T00:14:06.006083442Z" level=info msg=serving... address=/run/containerd/containerd.sock Nov 24 00:14:06.006475 containerd[1537]: time="2025-11-24T00:14:06.006330405Z" level=info msg="runtime interface starting up..." Nov 24 00:14:06.006475 containerd[1537]: time="2025-11-24T00:14:06.006417809Z" level=info msg="starting plugins..." Nov 24 00:14:06.006475 containerd[1537]: time="2025-11-24T00:14:06.006436283Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Nov 24 00:14:06.008679 containerd[1537]: time="2025-11-24T00:14:06.007652164Z" level=info msg="containerd successfully booted in 0.302668s" Nov 24 00:14:06.006788 systemd[1]: Started containerd.service - containerd container runtime. Nov 24 00:14:06.013410 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Nov 24 00:14:06.300861 systemd-networkd[1448]: eth0: Gained IPv6LL Nov 24 00:14:06.304252 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Nov 24 00:14:06.307125 systemd[1]: Reached target network-online.target - Network is Online. Nov 24 00:14:06.310547 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Nov 24 00:14:06.313716 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:14:06.332961 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Nov 24 00:14:06.361226 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Nov 24 00:14:06.364086 systemd[1]: coreos-metadata.service: Deactivated successfully. Nov 24 00:14:06.364349 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Nov 24 00:14:06.367955 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Nov 24 00:14:07.295772 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:14:07.298375 systemd[1]: Reached target multi-user.target - Multi-User System. Nov 24 00:14:07.300761 systemd[1]: Startup finished in 3.165s (kernel) + 8.498s (initrd) + 5.268s (userspace) = 16.932s. Nov 24 00:14:07.301053 (kubelet)[1639]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 24 00:14:07.737450 kubelet[1639]: E1124 00:14:07.737292 1639 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 24 00:14:07.741609 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 24 00:14:07.741864 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 24 00:14:07.742351 systemd[1]: kubelet.service: Consumed 1.219s CPU time, 263.9M memory peak. Nov 24 00:14:08.117533 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Nov 24 00:14:08.119201 systemd[1]: Started sshd@0-10.0.0.25:22-10.0.0.1:42088.service - OpenSSH per-connection server daemon (10.0.0.1:42088). Nov 24 00:14:08.200177 sshd[1652]: Accepted publickey for core from 10.0.0.1 port 42088 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:14:08.202385 sshd-session[1652]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:14:08.210784 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Nov 24 00:14:08.212278 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Nov 24 00:14:08.220986 systemd-logind[1519]: New session 1 of user core. Nov 24 00:14:08.243497 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Nov 24 00:14:08.246921 systemd[1]: Starting user@500.service - User Manager for UID 500... Nov 24 00:14:08.272168 (systemd)[1657]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Nov 24 00:14:08.275258 systemd-logind[1519]: New session c1 of user core. Nov 24 00:14:08.459325 systemd[1657]: Queued start job for default target default.target. Nov 24 00:14:08.480231 systemd[1657]: Created slice app.slice - User Application Slice. Nov 24 00:14:08.480265 systemd[1657]: Reached target paths.target - Paths. Nov 24 00:14:08.480323 systemd[1657]: Reached target timers.target - Timers. Nov 24 00:14:08.482088 systemd[1657]: Starting dbus.socket - D-Bus User Message Bus Socket... Nov 24 00:14:08.496303 systemd[1657]: Listening on dbus.socket - D-Bus User Message Bus Socket. Nov 24 00:14:08.496492 systemd[1657]: Reached target sockets.target - Sockets. Nov 24 00:14:08.496546 systemd[1657]: Reached target basic.target - Basic System. Nov 24 00:14:08.496587 systemd[1657]: Reached target default.target - Main User Target. Nov 24 00:14:08.496640 systemd[1657]: Startup finished in 214ms. Nov 24 00:14:08.497080 systemd[1]: Started user@500.service - User Manager for UID 500. Nov 24 00:14:08.498925 systemd[1]: Started session-1.scope - Session 1 of User core. Nov 24 00:14:08.564717 systemd[1]: Started sshd@1-10.0.0.25:22-10.0.0.1:42096.service - OpenSSH per-connection server daemon (10.0.0.1:42096). Nov 24 00:14:08.623479 sshd[1668]: Accepted publickey for core from 10.0.0.1 port 42096 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:14:08.625640 sshd-session[1668]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:14:08.631311 systemd-logind[1519]: New session 2 of user core. Nov 24 00:14:08.644996 systemd[1]: Started session-2.scope - Session 2 of User core. Nov 24 00:14:08.699417 sshd[1671]: Connection closed by 10.0.0.1 port 42096 Nov 24 00:14:08.700105 sshd-session[1668]: pam_unix(sshd:session): session closed for user core Nov 24 00:14:08.713848 systemd[1]: sshd@1-10.0.0.25:22-10.0.0.1:42096.service: Deactivated successfully. Nov 24 00:14:08.715920 systemd[1]: session-2.scope: Deactivated successfully. Nov 24 00:14:08.716746 systemd-logind[1519]: Session 2 logged out. Waiting for processes to exit. Nov 24 00:14:08.720131 systemd[1]: Started sshd@2-10.0.0.25:22-10.0.0.1:42098.service - OpenSSH per-connection server daemon (10.0.0.1:42098). Nov 24 00:14:08.720972 systemd-logind[1519]: Removed session 2. Nov 24 00:14:08.772148 sshd[1677]: Accepted publickey for core from 10.0.0.1 port 42098 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:14:08.773337 sshd-session[1677]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:14:08.778184 systemd-logind[1519]: New session 3 of user core. Nov 24 00:14:08.789839 systemd[1]: Started session-3.scope - Session 3 of User core. Nov 24 00:14:08.840507 sshd[1680]: Connection closed by 10.0.0.1 port 42098 Nov 24 00:14:08.840921 sshd-session[1677]: pam_unix(sshd:session): session closed for user core Nov 24 00:14:08.859855 systemd[1]: sshd@2-10.0.0.25:22-10.0.0.1:42098.service: Deactivated successfully. Nov 24 00:14:08.861671 systemd[1]: session-3.scope: Deactivated successfully. Nov 24 00:14:08.862450 systemd-logind[1519]: Session 3 logged out. Waiting for processes to exit. Nov 24 00:14:08.865091 systemd[1]: Started sshd@3-10.0.0.25:22-10.0.0.1:42108.service - OpenSSH per-connection server daemon (10.0.0.1:42108). Nov 24 00:14:08.865771 systemd-logind[1519]: Removed session 3. Nov 24 00:14:08.929125 sshd[1686]: Accepted publickey for core from 10.0.0.1 port 42108 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:14:08.930488 sshd-session[1686]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:14:08.935503 systemd-logind[1519]: New session 4 of user core. Nov 24 00:14:08.952956 systemd[1]: Started session-4.scope - Session 4 of User core. Nov 24 00:14:09.010157 sshd[1689]: Connection closed by 10.0.0.1 port 42108 Nov 24 00:14:09.010840 sshd-session[1686]: pam_unix(sshd:session): session closed for user core Nov 24 00:14:09.019992 systemd[1]: sshd@3-10.0.0.25:22-10.0.0.1:42108.service: Deactivated successfully. Nov 24 00:14:09.022016 systemd[1]: session-4.scope: Deactivated successfully. Nov 24 00:14:09.022782 systemd-logind[1519]: Session 4 logged out. Waiting for processes to exit. Nov 24 00:14:09.025730 systemd[1]: Started sshd@4-10.0.0.25:22-10.0.0.1:42118.service - OpenSSH per-connection server daemon (10.0.0.1:42118). Nov 24 00:14:09.026668 systemd-logind[1519]: Removed session 4. Nov 24 00:14:09.076985 sshd[1695]: Accepted publickey for core from 10.0.0.1 port 42118 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:14:09.078242 sshd-session[1695]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:14:09.083332 systemd-logind[1519]: New session 5 of user core. Nov 24 00:14:09.092795 systemd[1]: Started session-5.scope - Session 5 of User core. Nov 24 00:14:09.165292 sudo[1699]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Nov 24 00:14:09.165705 sudo[1699]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 24 00:14:09.188007 sudo[1699]: pam_unix(sudo:session): session closed for user root Nov 24 00:14:09.189704 sshd[1698]: Connection closed by 10.0.0.1 port 42118 Nov 24 00:14:09.190125 sshd-session[1695]: pam_unix(sshd:session): session closed for user core Nov 24 00:14:09.216984 systemd[1]: sshd@4-10.0.0.25:22-10.0.0.1:42118.service: Deactivated successfully. Nov 24 00:14:09.219035 systemd[1]: session-5.scope: Deactivated successfully. Nov 24 00:14:09.219935 systemd-logind[1519]: Session 5 logged out. Waiting for processes to exit. Nov 24 00:14:09.222808 systemd[1]: Started sshd@5-10.0.0.25:22-10.0.0.1:42122.service - OpenSSH per-connection server daemon (10.0.0.1:42122). Nov 24 00:14:09.223425 systemd-logind[1519]: Removed session 5. Nov 24 00:14:09.291388 sshd[1705]: Accepted publickey for core from 10.0.0.1 port 42122 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:14:09.292985 sshd-session[1705]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:14:09.297352 systemd-logind[1519]: New session 6 of user core. Nov 24 00:14:09.303890 systemd[1]: Started session-6.scope - Session 6 of User core. Nov 24 00:14:09.359307 sudo[1711]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Nov 24 00:14:09.359688 sudo[1711]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 24 00:14:09.417113 sudo[1711]: pam_unix(sudo:session): session closed for user root Nov 24 00:14:09.425868 sudo[1710]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Nov 24 00:14:09.426280 sudo[1710]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 24 00:14:09.438922 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 24 00:14:09.497811 augenrules[1733]: No rules Nov 24 00:14:09.499875 systemd[1]: audit-rules.service: Deactivated successfully. Nov 24 00:14:09.500726 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 24 00:14:09.502272 sudo[1710]: pam_unix(sudo:session): session closed for user root Nov 24 00:14:09.504348 sshd[1709]: Connection closed by 10.0.0.1 port 42122 Nov 24 00:14:09.504834 sshd-session[1705]: pam_unix(sshd:session): session closed for user core Nov 24 00:14:09.524794 systemd[1]: sshd@5-10.0.0.25:22-10.0.0.1:42122.service: Deactivated successfully. Nov 24 00:14:09.526873 systemd[1]: session-6.scope: Deactivated successfully. Nov 24 00:14:09.527696 systemd-logind[1519]: Session 6 logged out. Waiting for processes to exit. Nov 24 00:14:09.531054 systemd[1]: Started sshd@6-10.0.0.25:22-10.0.0.1:42130.service - OpenSSH per-connection server daemon (10.0.0.1:42130). Nov 24 00:14:09.531739 systemd-logind[1519]: Removed session 6. Nov 24 00:14:09.584455 sshd[1742]: Accepted publickey for core from 10.0.0.1 port 42130 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:14:09.586022 sshd-session[1742]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:14:09.590695 systemd-logind[1519]: New session 7 of user core. Nov 24 00:14:09.609900 systemd[1]: Started session-7.scope - Session 7 of User core. Nov 24 00:14:09.665258 sudo[1746]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Nov 24 00:14:09.665686 sudo[1746]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 24 00:14:10.016077 systemd[1]: Starting docker.service - Docker Application Container Engine... Nov 24 00:14:10.036177 (dockerd)[1767]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Nov 24 00:14:10.772742 dockerd[1767]: time="2025-11-24T00:14:10.772664853Z" level=info msg="Starting up" Nov 24 00:14:10.773663 dockerd[1767]: time="2025-11-24T00:14:10.773615676Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Nov 24 00:14:10.801250 dockerd[1767]: time="2025-11-24T00:14:10.801185620Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Nov 24 00:14:11.634991 dockerd[1767]: time="2025-11-24T00:14:11.634909871Z" level=info msg="Loading containers: start." Nov 24 00:14:11.646678 kernel: Initializing XFRM netlink socket Nov 24 00:14:11.937472 systemd-networkd[1448]: docker0: Link UP Nov 24 00:14:11.944292 dockerd[1767]: time="2025-11-24T00:14:11.944232711Z" level=info msg="Loading containers: done." Nov 24 00:14:11.978076 dockerd[1767]: time="2025-11-24T00:14:11.977993855Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Nov 24 00:14:11.978285 dockerd[1767]: time="2025-11-24T00:14:11.978158635Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Nov 24 00:14:11.978380 dockerd[1767]: time="2025-11-24T00:14:11.978339333Z" level=info msg="Initializing buildkit" Nov 24 00:14:12.029878 dockerd[1767]: time="2025-11-24T00:14:12.029804389Z" level=info msg="Completed buildkit initialization" Nov 24 00:14:12.034693 dockerd[1767]: time="2025-11-24T00:14:12.034643504Z" level=info msg="Daemon has completed initialization" Nov 24 00:14:12.034819 dockerd[1767]: time="2025-11-24T00:14:12.034760363Z" level=info msg="API listen on /run/docker.sock" Nov 24 00:14:12.034975 systemd[1]: Started docker.service - Docker Application Container Engine. Nov 24 00:14:12.941596 containerd[1537]: time="2025-11-24T00:14:12.941449609Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.10\"" Nov 24 00:14:13.689996 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1571891058.mount: Deactivated successfully. Nov 24 00:14:14.982921 containerd[1537]: time="2025-11-24T00:14:14.982817673Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:14.984076 containerd[1537]: time="2025-11-24T00:14:14.984001433Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.10: active requests=0, bytes read=29072183" Nov 24 00:14:14.986343 containerd[1537]: time="2025-11-24T00:14:14.986275738Z" level=info msg="ImageCreate event name:\"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:14.992177 containerd[1537]: time="2025-11-24T00:14:14.992092667Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:af4ee57c047e31a7f58422b94a9ec4c62221d3deebb16755bdeff720df796189\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:14.993250 containerd[1537]: time="2025-11-24T00:14:14.993195144Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.10\" with image id \"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:af4ee57c047e31a7f58422b94a9ec4c62221d3deebb16755bdeff720df796189\", size \"29068782\" in 2.051621633s" Nov 24 00:14:14.993321 containerd[1537]: time="2025-11-24T00:14:14.993252782Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.10\" returns image reference \"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\"" Nov 24 00:14:14.994214 containerd[1537]: time="2025-11-24T00:14:14.994164071Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.10\"" Nov 24 00:14:16.947111 containerd[1537]: time="2025-11-24T00:14:16.947013561Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:16.975306 containerd[1537]: time="2025-11-24T00:14:16.975209198Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.10: active requests=0, bytes read=24992010" Nov 24 00:14:17.012644 containerd[1537]: time="2025-11-24T00:14:17.012556598Z" level=info msg="ImageCreate event name:\"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:17.044161 containerd[1537]: time="2025-11-24T00:14:17.044076980Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:efbd9d1dfcd2940e1c73a1476c880c3c2cdf04cc60722d329b21cd48745c8660\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:17.045530 containerd[1537]: time="2025-11-24T00:14:17.045456797Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.10\" with image id \"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:efbd9d1dfcd2940e1c73a1476c880c3c2cdf04cc60722d329b21cd48745c8660\", size \"26649046\" in 2.051234858s" Nov 24 00:14:17.045530 containerd[1537]: time="2025-11-24T00:14:17.045520737Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.10\" returns image reference \"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\"" Nov 24 00:14:17.046335 containerd[1537]: time="2025-11-24T00:14:17.046292324Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.10\"" Nov 24 00:14:17.775266 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Nov 24 00:14:17.777437 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:14:18.064387 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:14:18.079334 (kubelet)[2057]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 24 00:14:18.174883 kubelet[2057]: E1124 00:14:18.174803 2057 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 24 00:14:18.182079 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 24 00:14:18.182455 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 24 00:14:18.183083 systemd[1]: kubelet.service: Consumed 311ms CPU time, 111.5M memory peak. Nov 24 00:14:19.011242 containerd[1537]: time="2025-11-24T00:14:19.011169450Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:19.012011 containerd[1537]: time="2025-11-24T00:14:19.011969911Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.10: active requests=0, bytes read=19404248" Nov 24 00:14:19.013291 containerd[1537]: time="2025-11-24T00:14:19.013235996Z" level=info msg="ImageCreate event name:\"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:19.016165 containerd[1537]: time="2025-11-24T00:14:19.016104135Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:9c58e1adcad5af66d1d9ca5cf9a4c266e4054b8f19f91a8fff1993549e657b10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:19.017102 containerd[1537]: time="2025-11-24T00:14:19.017063223Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.10\" with image id \"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:9c58e1adcad5af66d1d9ca5cf9a4c266e4054b8f19f91a8fff1993549e657b10\", size \"21061302\" in 1.970735863s" Nov 24 00:14:19.017102 containerd[1537]: time="2025-11-24T00:14:19.017097898Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.10\" returns image reference \"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\"" Nov 24 00:14:19.017648 containerd[1537]: time="2025-11-24T00:14:19.017557290Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.10\"" Nov 24 00:14:20.578279 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2441738879.mount: Deactivated successfully. Nov 24 00:14:21.498045 containerd[1537]: time="2025-11-24T00:14:21.497942317Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:21.498952 containerd[1537]: time="2025-11-24T00:14:21.498909310Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.10: active requests=0, bytes read=31161423" Nov 24 00:14:21.500838 containerd[1537]: time="2025-11-24T00:14:21.500782774Z" level=info msg="ImageCreate event name:\"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:21.504065 containerd[1537]: time="2025-11-24T00:14:21.503990229Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e3dda1c7b384f9eb5b2fa1c27493b23b80e6204b9fa2ee8791b2de078f468cbf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:21.505511 containerd[1537]: time="2025-11-24T00:14:21.505280559Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.10\" with image id \"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\", repo tag \"registry.k8s.io/kube-proxy:v1.32.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:e3dda1c7b384f9eb5b2fa1c27493b23b80e6204b9fa2ee8791b2de078f468cbf\", size \"31160442\" in 2.487684867s" Nov 24 00:14:21.505511 containerd[1537]: time="2025-11-24T00:14:21.505339089Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.10\" returns image reference \"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\"" Nov 24 00:14:21.506416 containerd[1537]: time="2025-11-24T00:14:21.506102450Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Nov 24 00:14:22.158518 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3063843717.mount: Deactivated successfully. Nov 24 00:14:23.862522 containerd[1537]: time="2025-11-24T00:14:23.860990796Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:23.893819 containerd[1537]: time="2025-11-24T00:14:23.893159945Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Nov 24 00:14:23.927538 containerd[1537]: time="2025-11-24T00:14:23.927444210Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:23.995027 containerd[1537]: time="2025-11-24T00:14:23.994913369Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:23.996260 containerd[1537]: time="2025-11-24T00:14:23.996202386Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 2.490062366s" Nov 24 00:14:23.996260 containerd[1537]: time="2025-11-24T00:14:23.996255065Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Nov 24 00:14:23.996994 containerd[1537]: time="2025-11-24T00:14:23.996835584Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Nov 24 00:14:24.709305 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4075828187.mount: Deactivated successfully. Nov 24 00:14:24.717152 containerd[1537]: time="2025-11-24T00:14:24.717090769Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 24 00:14:24.718203 containerd[1537]: time="2025-11-24T00:14:24.718131020Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Nov 24 00:14:24.721092 containerd[1537]: time="2025-11-24T00:14:24.721013376Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 24 00:14:24.724230 containerd[1537]: time="2025-11-24T00:14:24.724171458Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 24 00:14:24.725130 containerd[1537]: time="2025-11-24T00:14:24.725077507Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 728.198412ms" Nov 24 00:14:24.725130 containerd[1537]: time="2025-11-24T00:14:24.725114777Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Nov 24 00:14:24.727069 containerd[1537]: time="2025-11-24T00:14:24.725857059Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Nov 24 00:14:26.032251 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4110432896.mount: Deactivated successfully. Nov 24 00:14:28.275344 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Nov 24 00:14:28.277690 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:14:28.536194 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:14:28.555956 (kubelet)[2166]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 24 00:14:28.749352 kubelet[2166]: E1124 00:14:28.749217 2166 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 24 00:14:28.753054 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 24 00:14:28.753296 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 24 00:14:28.753719 systemd[1]: kubelet.service: Consumed 268ms CPU time, 111M memory peak. Nov 24 00:14:34.360404 containerd[1537]: time="2025-11-24T00:14:34.360313464Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:34.363727 containerd[1537]: time="2025-11-24T00:14:34.363667244Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Nov 24 00:14:34.370369 containerd[1537]: time="2025-11-24T00:14:34.370307226Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:34.377326 containerd[1537]: time="2025-11-24T00:14:34.377265976Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:34.378767 containerd[1537]: time="2025-11-24T00:14:34.378703883Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 9.651881122s" Nov 24 00:14:34.378767 containerd[1537]: time="2025-11-24T00:14:34.378753105Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Nov 24 00:14:36.726761 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:14:36.727023 systemd[1]: kubelet.service: Consumed 268ms CPU time, 111M memory peak. Nov 24 00:14:36.729482 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:14:36.755352 systemd[1]: Reload requested from client PID 2234 ('systemctl') (unit session-7.scope)... Nov 24 00:14:36.755369 systemd[1]: Reloading... Nov 24 00:14:36.887741 zram_generator::config[2279]: No configuration found. Nov 24 00:14:37.146385 systemd[1]: Reloading finished in 390 ms. Nov 24 00:14:37.227889 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Nov 24 00:14:37.228043 systemd[1]: kubelet.service: Failed with result 'signal'. Nov 24 00:14:37.228390 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:14:37.228454 systemd[1]: kubelet.service: Consumed 192ms CPU time, 98.2M memory peak. Nov 24 00:14:37.230298 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:14:37.431448 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:14:37.435878 (kubelet)[2324]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 24 00:14:37.502156 kubelet[2324]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:14:37.502156 kubelet[2324]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 24 00:14:37.502156 kubelet[2324]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:14:37.502780 kubelet[2324]: I1124 00:14:37.502241 2324 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 00:14:37.977566 kubelet[2324]: I1124 00:14:37.977490 2324 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 24 00:14:37.977566 kubelet[2324]: I1124 00:14:37.977536 2324 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 00:14:37.977953 kubelet[2324]: I1124 00:14:37.977918 2324 server.go:954] "Client rotation is on, will bootstrap in background" Nov 24 00:14:38.009303 kubelet[2324]: E1124 00:14:38.009228 2324 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:38.010154 kubelet[2324]: I1124 00:14:38.010120 2324 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 24 00:14:38.026458 kubelet[2324]: I1124 00:14:38.026402 2324 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 00:14:38.033556 kubelet[2324]: I1124 00:14:38.033496 2324 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 24 00:14:38.035896 kubelet[2324]: I1124 00:14:38.035810 2324 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 00:14:38.036137 kubelet[2324]: I1124 00:14:38.035876 2324 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 00:14:38.036137 kubelet[2324]: I1124 00:14:38.036129 2324 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 00:14:38.036137 kubelet[2324]: I1124 00:14:38.036140 2324 container_manager_linux.go:304] "Creating device plugin manager" Nov 24 00:14:38.036477 kubelet[2324]: I1124 00:14:38.036323 2324 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:14:38.045662 kubelet[2324]: I1124 00:14:38.045545 2324 kubelet.go:446] "Attempting to sync node with API server" Nov 24 00:14:38.045851 kubelet[2324]: I1124 00:14:38.045702 2324 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 00:14:38.045851 kubelet[2324]: I1124 00:14:38.045744 2324 kubelet.go:352] "Adding apiserver pod source" Nov 24 00:14:38.045851 kubelet[2324]: I1124 00:14:38.045761 2324 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 00:14:38.048060 kubelet[2324]: W1124 00:14:38.047978 2324 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.25:6443: connect: connection refused Nov 24 00:14:38.048060 kubelet[2324]: W1124 00:14:38.048020 2324 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.25:6443: connect: connection refused Nov 24 00:14:38.048681 kubelet[2324]: E1124 00:14:38.048066 2324 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:38.048681 kubelet[2324]: E1124 00:14:38.048108 2324 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:38.050876 kubelet[2324]: I1124 00:14:38.050838 2324 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Nov 24 00:14:38.051371 kubelet[2324]: I1124 00:14:38.051339 2324 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 00:14:38.054465 kubelet[2324]: W1124 00:14:38.051458 2324 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Nov 24 00:14:38.055706 kubelet[2324]: I1124 00:14:38.055664 2324 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 24 00:14:38.055760 kubelet[2324]: I1124 00:14:38.055725 2324 server.go:1287] "Started kubelet" Nov 24 00:14:38.055976 kubelet[2324]: I1124 00:14:38.055942 2324 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 00:14:38.057679 kubelet[2324]: I1124 00:14:38.057643 2324 server.go:479] "Adding debug handlers to kubelet server" Nov 24 00:14:38.059007 kubelet[2324]: I1124 00:14:38.058936 2324 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 00:14:38.060318 kubelet[2324]: I1124 00:14:38.059247 2324 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 00:14:38.060318 kubelet[2324]: I1124 00:14:38.059284 2324 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 00:14:38.060318 kubelet[2324]: I1124 00:14:38.059587 2324 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 24 00:14:38.061979 kubelet[2324]: E1124 00:14:38.061936 2324 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 24 00:14:38.062141 kubelet[2324]: I1124 00:14:38.062008 2324 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 24 00:14:38.063651 kubelet[2324]: E1124 00:14:38.062378 2324 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="200ms" Nov 24 00:14:38.063651 kubelet[2324]: I1124 00:14:38.062580 2324 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 24 00:14:38.063651 kubelet[2324]: I1124 00:14:38.062658 2324 reconciler.go:26] "Reconciler: start to sync state" Nov 24 00:14:38.063651 kubelet[2324]: W1124 00:14:38.063024 2324 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.25:6443: connect: connection refused Nov 24 00:14:38.063651 kubelet[2324]: E1124 00:14:38.063056 2324 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:38.067803 kubelet[2324]: E1124 00:14:38.063311 2324 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.25:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.187ac910cc728a6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-11-24 00:14:38.055688811 +0000 UTC m=+0.612777366,LastTimestamp:2025-11-24 00:14:38.055688811 +0000 UTC m=+0.612777366,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Nov 24 00:14:38.068913 kubelet[2324]: E1124 00:14:38.068750 2324 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 24 00:14:38.068913 kubelet[2324]: I1124 00:14:38.068779 2324 factory.go:221] Registration of the containerd container factory successfully Nov 24 00:14:38.068913 kubelet[2324]: I1124 00:14:38.068826 2324 factory.go:221] Registration of the systemd container factory successfully Nov 24 00:14:38.069072 kubelet[2324]: I1124 00:14:38.069035 2324 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 24 00:14:38.085580 kubelet[2324]: I1124 00:14:38.085534 2324 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 24 00:14:38.085580 kubelet[2324]: I1124 00:14:38.085562 2324 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 24 00:14:38.085580 kubelet[2324]: I1124 00:14:38.085582 2324 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:14:38.089509 kubelet[2324]: I1124 00:14:38.089448 2324 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 00:14:38.091728 kubelet[2324]: I1124 00:14:38.091700 2324 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 00:14:38.091728 kubelet[2324]: I1124 00:14:38.091721 2324 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 24 00:14:38.091827 kubelet[2324]: I1124 00:14:38.091742 2324 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 24 00:14:38.091827 kubelet[2324]: I1124 00:14:38.091750 2324 kubelet.go:2382] "Starting kubelet main sync loop" Nov 24 00:14:38.091827 kubelet[2324]: E1124 00:14:38.091798 2324 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 00:14:38.093690 kubelet[2324]: W1124 00:14:38.092317 2324 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.25:6443: connect: connection refused Nov 24 00:14:38.093690 kubelet[2324]: E1124 00:14:38.092359 2324 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:38.162779 kubelet[2324]: E1124 00:14:38.162698 2324 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 24 00:14:38.192130 kubelet[2324]: E1124 00:14:38.192049 2324 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 24 00:14:38.263128 kubelet[2324]: E1124 00:14:38.262960 2324 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 24 00:14:38.263128 kubelet[2324]: E1124 00:14:38.263078 2324 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="400ms" Nov 24 00:14:38.363614 kubelet[2324]: E1124 00:14:38.363537 2324 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 24 00:14:38.392848 kubelet[2324]: E1124 00:14:38.392756 2324 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 24 00:14:38.464505 kubelet[2324]: E1124 00:14:38.464382 2324 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 24 00:14:38.502744 kubelet[2324]: I1124 00:14:38.502674 2324 policy_none.go:49] "None policy: Start" Nov 24 00:14:38.502744 kubelet[2324]: I1124 00:14:38.502720 2324 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 24 00:14:38.502744 kubelet[2324]: I1124 00:14:38.502736 2324 state_mem.go:35] "Initializing new in-memory state store" Nov 24 00:14:38.516108 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Nov 24 00:14:38.531383 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Nov 24 00:14:38.535342 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Nov 24 00:14:38.554855 kubelet[2324]: I1124 00:14:38.554818 2324 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 00:14:38.555131 kubelet[2324]: I1124 00:14:38.555102 2324 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 00:14:38.555236 kubelet[2324]: I1124 00:14:38.555123 2324 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 00:14:38.555375 kubelet[2324]: I1124 00:14:38.555361 2324 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 00:14:38.556325 kubelet[2324]: E1124 00:14:38.556300 2324 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 24 00:14:38.556395 kubelet[2324]: E1124 00:14:38.556333 2324 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Nov 24 00:14:38.608482 kubelet[2324]: E1124 00:14:38.608348 2324 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.25:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.187ac910cc728a6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-11-24 00:14:38.055688811 +0000 UTC m=+0.612777366,LastTimestamp:2025-11-24 00:14:38.055688811 +0000 UTC m=+0.612777366,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Nov 24 00:14:38.656829 kubelet[2324]: I1124 00:14:38.656771 2324 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 24 00:14:38.657210 kubelet[2324]: E1124 00:14:38.657152 2324 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Nov 24 00:14:38.663778 kubelet[2324]: E1124 00:14:38.663727 2324 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="800ms" Nov 24 00:14:38.802817 systemd[1]: Created slice kubepods-burstable-poddb6cd6cee75c80f34a5bc34c3c506b23.slice - libcontainer container kubepods-burstable-poddb6cd6cee75c80f34a5bc34c3c506b23.slice. Nov 24 00:14:38.817918 kubelet[2324]: E1124 00:14:38.817861 2324 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 24 00:14:38.822008 systemd[1]: Created slice kubepods-burstable-pod55d9ac750f8c9141f337af8b08cf5c9d.slice - libcontainer container kubepods-burstable-pod55d9ac750f8c9141f337af8b08cf5c9d.slice. Nov 24 00:14:38.824091 kubelet[2324]: E1124 00:14:38.824059 2324 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 24 00:14:38.826804 systemd[1]: Created slice kubepods-burstable-pod0a68423804124305a9de061f38780871.slice - libcontainer container kubepods-burstable-pod0a68423804124305a9de061f38780871.slice. Nov 24 00:14:38.828396 kubelet[2324]: E1124 00:14:38.828358 2324 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 24 00:14:38.858752 kubelet[2324]: I1124 00:14:38.858713 2324 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 24 00:14:38.859053 kubelet[2324]: E1124 00:14:38.859032 2324 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Nov 24 00:14:38.866614 kubelet[2324]: I1124 00:14:38.866566 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0a68423804124305a9de061f38780871-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"0a68423804124305a9de061f38780871\") " pod="kube-system/kube-scheduler-localhost" Nov 24 00:14:38.866614 kubelet[2324]: I1124 00:14:38.866600 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/db6cd6cee75c80f34a5bc34c3c506b23-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"db6cd6cee75c80f34a5bc34c3c506b23\") " pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:38.866734 kubelet[2324]: I1124 00:14:38.866647 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/db6cd6cee75c80f34a5bc34c3c506b23-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"db6cd6cee75c80f34a5bc34c3c506b23\") " pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:38.866734 kubelet[2324]: I1124 00:14:38.866678 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:38.866734 kubelet[2324]: I1124 00:14:38.866699 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:38.866849 kubelet[2324]: I1124 00:14:38.866788 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:38.866895 kubelet[2324]: I1124 00:14:38.866860 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/db6cd6cee75c80f34a5bc34c3c506b23-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"db6cd6cee75c80f34a5bc34c3c506b23\") " pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:38.866895 kubelet[2324]: I1124 00:14:38.866880 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:38.866969 kubelet[2324]: I1124 00:14:38.866915 2324 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:38.957363 kubelet[2324]: W1124 00:14:38.957283 2324 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.25:6443: connect: connection refused Nov 24 00:14:38.957363 kubelet[2324]: E1124 00:14:38.957346 2324 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:39.068405 kubelet[2324]: W1124 00:14:39.068221 2324 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.25:6443: connect: connection refused Nov 24 00:14:39.068405 kubelet[2324]: E1124 00:14:39.068304 2324 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:39.118890 kubelet[2324]: E1124 00:14:39.118814 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:39.119791 containerd[1537]: time="2025-11-24T00:14:39.119725551Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:db6cd6cee75c80f34a5bc34c3c506b23,Namespace:kube-system,Attempt:0,}" Nov 24 00:14:39.124983 kubelet[2324]: E1124 00:14:39.124954 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:39.125584 containerd[1537]: time="2025-11-24T00:14:39.125520047Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:55d9ac750f8c9141f337af8b08cf5c9d,Namespace:kube-system,Attempt:0,}" Nov 24 00:14:39.129808 kubelet[2324]: E1124 00:14:39.129768 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:39.130349 containerd[1537]: time="2025-11-24T00:14:39.130298641Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:0a68423804124305a9de061f38780871,Namespace:kube-system,Attempt:0,}" Nov 24 00:14:39.136916 kubelet[2324]: W1124 00:14:39.136874 2324 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.25:6443: connect: connection refused Nov 24 00:14:39.136961 kubelet[2324]: E1124 00:14:39.136926 2324 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:39.261227 kubelet[2324]: I1124 00:14:39.261186 2324 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 24 00:14:39.261618 kubelet[2324]: E1124 00:14:39.261583 2324 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Nov 24 00:14:39.434831 kubelet[2324]: W1124 00:14:39.434645 2324 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.25:6443: connect: connection refused Nov 24 00:14:39.434831 kubelet[2324]: E1124 00:14:39.434725 2324 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" Nov 24 00:14:39.464977 kubelet[2324]: E1124 00:14:39.464918 2324 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="1.6s" Nov 24 00:14:39.526183 containerd[1537]: time="2025-11-24T00:14:39.526126639Z" level=info msg="connecting to shim db2a9802b6c18a6015996ff4df1b15f7e013bb67e49b707fa20a4632a2251d4c" address="unix:///run/containerd/s/bb56b4d7cff28f82b91cbc48ee2eaaa291ac1cbb490a3fef71c41fa892ad3976" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:14:39.542818 containerd[1537]: time="2025-11-24T00:14:39.542753964Z" level=info msg="connecting to shim 6db402dbb13affba994f5237709989e18b0a2ca715eb52b2925aea3dc046dd68" address="unix:///run/containerd/s/cd486fb0aedb5dc68833a26595544611e94c384e6c25d6672770094cb0131e94" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:14:39.553144 containerd[1537]: time="2025-11-24T00:14:39.553046855Z" level=info msg="connecting to shim f968c0bd963844765b00c4e65ccc8f650ad733c0a162882e8308ba9a7de97a8c" address="unix:///run/containerd/s/f91ade52eac2514186d213830e3dd051e2319c2009a268ceffc08c8031772356" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:14:39.618987 systemd[1]: Started cri-containerd-db2a9802b6c18a6015996ff4df1b15f7e013bb67e49b707fa20a4632a2251d4c.scope - libcontainer container db2a9802b6c18a6015996ff4df1b15f7e013bb67e49b707fa20a4632a2251d4c. Nov 24 00:14:39.627179 systemd[1]: Started cri-containerd-6db402dbb13affba994f5237709989e18b0a2ca715eb52b2925aea3dc046dd68.scope - libcontainer container 6db402dbb13affba994f5237709989e18b0a2ca715eb52b2925aea3dc046dd68. Nov 24 00:14:39.629378 systemd[1]: Started cri-containerd-f968c0bd963844765b00c4e65ccc8f650ad733c0a162882e8308ba9a7de97a8c.scope - libcontainer container f968c0bd963844765b00c4e65ccc8f650ad733c0a162882e8308ba9a7de97a8c. Nov 24 00:14:39.753199 containerd[1537]: time="2025-11-24T00:14:39.753051767Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:db6cd6cee75c80f34a5bc34c3c506b23,Namespace:kube-system,Attempt:0,} returns sandbox id \"f968c0bd963844765b00c4e65ccc8f650ad733c0a162882e8308ba9a7de97a8c\"" Nov 24 00:14:39.753585 containerd[1537]: time="2025-11-24T00:14:39.753559272Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:0a68423804124305a9de061f38780871,Namespace:kube-system,Attempt:0,} returns sandbox id \"6db402dbb13affba994f5237709989e18b0a2ca715eb52b2925aea3dc046dd68\"" Nov 24 00:14:39.754974 containerd[1537]: time="2025-11-24T00:14:39.754924135Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:55d9ac750f8c9141f337af8b08cf5c9d,Namespace:kube-system,Attempt:0,} returns sandbox id \"db2a9802b6c18a6015996ff4df1b15f7e013bb67e49b707fa20a4632a2251d4c\"" Nov 24 00:14:39.755388 kubelet[2324]: E1124 00:14:39.755359 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:39.755792 kubelet[2324]: E1124 00:14:39.755592 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:39.756783 kubelet[2324]: E1124 00:14:39.756757 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:39.758557 containerd[1537]: time="2025-11-24T00:14:39.758396118Z" level=info msg="CreateContainer within sandbox \"6db402dbb13affba994f5237709989e18b0a2ca715eb52b2925aea3dc046dd68\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Nov 24 00:14:39.758557 containerd[1537]: time="2025-11-24T00:14:39.758441976Z" level=info msg="CreateContainer within sandbox \"db2a9802b6c18a6015996ff4df1b15f7e013bb67e49b707fa20a4632a2251d4c\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Nov 24 00:14:39.758677 containerd[1537]: time="2025-11-24T00:14:39.758576294Z" level=info msg="CreateContainer within sandbox \"f968c0bd963844765b00c4e65ccc8f650ad733c0a162882e8308ba9a7de97a8c\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Nov 24 00:14:39.778581 containerd[1537]: time="2025-11-24T00:14:39.778511756Z" level=info msg="Container 122c3278f133f41ed55e282836f6b036044ddd95857cdf190b4ff8ee996f2ae2: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:14:39.781757 containerd[1537]: time="2025-11-24T00:14:39.781699352Z" level=info msg="Container 47a963265348336e642d905319da1d28f2fc59baebba408be52cd68e59dc0b5f: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:14:39.785129 containerd[1537]: time="2025-11-24T00:14:39.785078306Z" level=info msg="Container c1c7609bf40f66390e0ab1a8cfd09c2157e6b57e7111b0b8d2562ed3f1c0a75d: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:14:39.789834 containerd[1537]: time="2025-11-24T00:14:39.789770133Z" level=info msg="CreateContainer within sandbox \"6db402dbb13affba994f5237709989e18b0a2ca715eb52b2925aea3dc046dd68\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"122c3278f133f41ed55e282836f6b036044ddd95857cdf190b4ff8ee996f2ae2\"" Nov 24 00:14:39.790551 containerd[1537]: time="2025-11-24T00:14:39.790525494Z" level=info msg="StartContainer for \"122c3278f133f41ed55e282836f6b036044ddd95857cdf190b4ff8ee996f2ae2\"" Nov 24 00:14:39.791681 containerd[1537]: time="2025-11-24T00:14:39.791652941Z" level=info msg="connecting to shim 122c3278f133f41ed55e282836f6b036044ddd95857cdf190b4ff8ee996f2ae2" address="unix:///run/containerd/s/cd486fb0aedb5dc68833a26595544611e94c384e6c25d6672770094cb0131e94" protocol=ttrpc version=3 Nov 24 00:14:39.801211 containerd[1537]: time="2025-11-24T00:14:39.801145714Z" level=info msg="CreateContainer within sandbox \"f968c0bd963844765b00c4e65ccc8f650ad733c0a162882e8308ba9a7de97a8c\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"c1c7609bf40f66390e0ab1a8cfd09c2157e6b57e7111b0b8d2562ed3f1c0a75d\"" Nov 24 00:14:39.801697 containerd[1537]: time="2025-11-24T00:14:39.801647609Z" level=info msg="CreateContainer within sandbox \"db2a9802b6c18a6015996ff4df1b15f7e013bb67e49b707fa20a4632a2251d4c\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"47a963265348336e642d905319da1d28f2fc59baebba408be52cd68e59dc0b5f\"" Nov 24 00:14:39.801958 containerd[1537]: time="2025-11-24T00:14:39.801770435Z" level=info msg="StartContainer for \"c1c7609bf40f66390e0ab1a8cfd09c2157e6b57e7111b0b8d2562ed3f1c0a75d\"" Nov 24 00:14:39.802333 containerd[1537]: time="2025-11-24T00:14:39.802295484Z" level=info msg="StartContainer for \"47a963265348336e642d905319da1d28f2fc59baebba408be52cd68e59dc0b5f\"" Nov 24 00:14:39.803617 containerd[1537]: time="2025-11-24T00:14:39.803582327Z" level=info msg="connecting to shim c1c7609bf40f66390e0ab1a8cfd09c2157e6b57e7111b0b8d2562ed3f1c0a75d" address="unix:///run/containerd/s/f91ade52eac2514186d213830e3dd051e2319c2009a268ceffc08c8031772356" protocol=ttrpc version=3 Nov 24 00:14:39.804854 containerd[1537]: time="2025-11-24T00:14:39.804808373Z" level=info msg="connecting to shim 47a963265348336e642d905319da1d28f2fc59baebba408be52cd68e59dc0b5f" address="unix:///run/containerd/s/bb56b4d7cff28f82b91cbc48ee2eaaa291ac1cbb490a3fef71c41fa892ad3976" protocol=ttrpc version=3 Nov 24 00:14:39.816508 systemd[1]: Started cri-containerd-122c3278f133f41ed55e282836f6b036044ddd95857cdf190b4ff8ee996f2ae2.scope - libcontainer container 122c3278f133f41ed55e282836f6b036044ddd95857cdf190b4ff8ee996f2ae2. Nov 24 00:14:39.847995 systemd[1]: Started cri-containerd-47a963265348336e642d905319da1d28f2fc59baebba408be52cd68e59dc0b5f.scope - libcontainer container 47a963265348336e642d905319da1d28f2fc59baebba408be52cd68e59dc0b5f. Nov 24 00:14:39.863853 systemd[1]: Started cri-containerd-c1c7609bf40f66390e0ab1a8cfd09c2157e6b57e7111b0b8d2562ed3f1c0a75d.scope - libcontainer container c1c7609bf40f66390e0ab1a8cfd09c2157e6b57e7111b0b8d2562ed3f1c0a75d. Nov 24 00:14:39.915245 containerd[1537]: time="2025-11-24T00:14:39.915169958Z" level=info msg="StartContainer for \"122c3278f133f41ed55e282836f6b036044ddd95857cdf190b4ff8ee996f2ae2\" returns successfully" Nov 24 00:14:39.939045 containerd[1537]: time="2025-11-24T00:14:39.938872670Z" level=info msg="StartContainer for \"c1c7609bf40f66390e0ab1a8cfd09c2157e6b57e7111b0b8d2562ed3f1c0a75d\" returns successfully" Nov 24 00:14:39.940661 containerd[1537]: time="2025-11-24T00:14:39.940505408Z" level=info msg="StartContainer for \"47a963265348336e642d905319da1d28f2fc59baebba408be52cd68e59dc0b5f\" returns successfully" Nov 24 00:14:40.063074 kubelet[2324]: I1124 00:14:40.063026 2324 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 24 00:14:40.103049 kubelet[2324]: E1124 00:14:40.102992 2324 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 24 00:14:40.103338 kubelet[2324]: E1124 00:14:40.103132 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:40.109040 kubelet[2324]: E1124 00:14:40.109001 2324 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 24 00:14:40.109167 kubelet[2324]: E1124 00:14:40.109140 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:40.110450 kubelet[2324]: E1124 00:14:40.110424 2324 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 24 00:14:40.110549 kubelet[2324]: E1124 00:14:40.110526 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:41.113007 kubelet[2324]: E1124 00:14:41.112963 2324 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 24 00:14:41.113452 kubelet[2324]: E1124 00:14:41.113101 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:41.113810 kubelet[2324]: E1124 00:14:41.113779 2324 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 24 00:14:41.113911 kubelet[2324]: E1124 00:14:41.113893 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:41.182998 kubelet[2324]: E1124 00:14:41.182929 2324 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Nov 24 00:14:41.255916 kubelet[2324]: I1124 00:14:41.255826 2324 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Nov 24 00:14:41.262985 kubelet[2324]: I1124 00:14:41.262922 2324 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:41.271457 kubelet[2324]: E1124 00:14:41.271405 2324 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:41.271457 kubelet[2324]: I1124 00:14:41.271438 2324 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:41.273768 kubelet[2324]: E1124 00:14:41.273738 2324 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:41.273768 kubelet[2324]: I1124 00:14:41.273765 2324 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Nov 24 00:14:41.275571 kubelet[2324]: E1124 00:14:41.275534 2324 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Nov 24 00:14:42.050521 kubelet[2324]: I1124 00:14:42.050411 2324 apiserver.go:52] "Watching apiserver" Nov 24 00:14:42.063731 kubelet[2324]: I1124 00:14:42.063656 2324 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 24 00:14:43.426415 kubelet[2324]: I1124 00:14:43.426368 2324 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:43.465228 kubelet[2324]: E1124 00:14:43.465078 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:43.966605 systemd[1]: Reload requested from client PID 2595 ('systemctl') (unit session-7.scope)... Nov 24 00:14:43.966648 systemd[1]: Reloading... Nov 24 00:14:44.072685 zram_generator::config[2641]: No configuration found. Nov 24 00:14:44.116858 kubelet[2324]: E1124 00:14:44.116805 2324 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:44.411479 systemd[1]: Reloading finished in 444 ms. Nov 24 00:14:44.451308 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:14:44.468466 systemd[1]: kubelet.service: Deactivated successfully. Nov 24 00:14:44.469016 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:14:44.469112 systemd[1]: kubelet.service: Consumed 1.156s CPU time, 134M memory peak. Nov 24 00:14:44.471985 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 24 00:14:44.824265 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 24 00:14:44.841722 (kubelet)[2683]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 24 00:14:45.134738 kubelet[2683]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:14:45.134738 kubelet[2683]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 24 00:14:45.134738 kubelet[2683]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 00:14:45.134738 kubelet[2683]: I1124 00:14:45.134650 2683 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 00:14:45.142421 kubelet[2683]: I1124 00:14:45.142379 2683 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 24 00:14:45.142421 kubelet[2683]: I1124 00:14:45.142402 2683 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 00:14:45.142703 kubelet[2683]: I1124 00:14:45.142676 2683 server.go:954] "Client rotation is on, will bootstrap in background" Nov 24 00:14:45.143923 kubelet[2683]: I1124 00:14:45.143891 2683 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 00:14:45.146277 kubelet[2683]: I1124 00:14:45.146222 2683 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 24 00:14:45.151942 kubelet[2683]: I1124 00:14:45.151913 2683 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 00:14:45.158012 kubelet[2683]: I1124 00:14:45.157976 2683 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 24 00:14:45.158322 kubelet[2683]: I1124 00:14:45.158272 2683 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 00:14:45.158542 kubelet[2683]: I1124 00:14:45.158311 2683 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 00:14:45.158661 kubelet[2683]: I1124 00:14:45.158558 2683 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 00:14:45.158661 kubelet[2683]: I1124 00:14:45.158571 2683 container_manager_linux.go:304] "Creating device plugin manager" Nov 24 00:14:45.158716 kubelet[2683]: I1124 00:14:45.158662 2683 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:14:45.158920 kubelet[2683]: I1124 00:14:45.158893 2683 kubelet.go:446] "Attempting to sync node with API server" Nov 24 00:14:45.158946 kubelet[2683]: I1124 00:14:45.158926 2683 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 00:14:45.158972 kubelet[2683]: I1124 00:14:45.158954 2683 kubelet.go:352] "Adding apiserver pod source" Nov 24 00:14:45.158972 kubelet[2683]: I1124 00:14:45.158968 2683 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 00:14:45.161816 kubelet[2683]: I1124 00:14:45.161778 2683 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Nov 24 00:14:45.162910 kubelet[2683]: I1124 00:14:45.162892 2683 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 00:14:45.163538 kubelet[2683]: I1124 00:14:45.163509 2683 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 24 00:14:45.163683 kubelet[2683]: I1124 00:14:45.163667 2683 server.go:1287] "Started kubelet" Nov 24 00:14:45.166063 kubelet[2683]: I1124 00:14:45.165977 2683 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 00:14:45.166358 kubelet[2683]: I1124 00:14:45.166201 2683 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 00:14:45.169665 kubelet[2683]: I1124 00:14:45.166694 2683 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 00:14:45.169665 kubelet[2683]: I1124 00:14:45.168849 2683 server.go:479] "Adding debug handlers to kubelet server" Nov 24 00:14:45.170166 kubelet[2683]: I1124 00:14:45.170137 2683 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 00:14:45.171615 kubelet[2683]: I1124 00:14:45.171571 2683 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 24 00:14:45.172094 kubelet[2683]: I1124 00:14:45.172070 2683 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 24 00:14:45.173861 kubelet[2683]: I1124 00:14:45.173838 2683 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 24 00:14:45.174243 kubelet[2683]: I1124 00:14:45.174031 2683 reconciler.go:26] "Reconciler: start to sync state" Nov 24 00:14:45.176261 kubelet[2683]: E1124 00:14:45.176236 2683 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 24 00:14:45.179172 kubelet[2683]: I1124 00:14:45.179136 2683 factory.go:221] Registration of the systemd container factory successfully Nov 24 00:14:45.179238 kubelet[2683]: I1124 00:14:45.179223 2683 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 24 00:14:45.180307 kubelet[2683]: I1124 00:14:45.180269 2683 factory.go:221] Registration of the containerd container factory successfully Nov 24 00:14:45.187661 kubelet[2683]: I1124 00:14:45.187610 2683 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 00:14:45.189216 kubelet[2683]: I1124 00:14:45.189201 2683 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 00:14:45.189292 kubelet[2683]: I1124 00:14:45.189283 2683 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 24 00:14:45.189397 kubelet[2683]: I1124 00:14:45.189376 2683 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 24 00:14:45.189445 kubelet[2683]: I1124 00:14:45.189437 2683 kubelet.go:2382] "Starting kubelet main sync loop" Nov 24 00:14:45.189588 kubelet[2683]: E1124 00:14:45.189549 2683 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 00:14:45.223731 kubelet[2683]: I1124 00:14:45.223686 2683 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 24 00:14:45.223993 kubelet[2683]: I1124 00:14:45.223933 2683 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 24 00:14:45.223993 kubelet[2683]: I1124 00:14:45.223984 2683 state_mem.go:36] "Initialized new in-memory state store" Nov 24 00:14:45.224199 kubelet[2683]: I1124 00:14:45.224165 2683 state_mem.go:88] "Updated default CPUSet" cpuSet="" Nov 24 00:14:45.224199 kubelet[2683]: I1124 00:14:45.224176 2683 state_mem.go:96] "Updated CPUSet assignments" assignments={} Nov 24 00:14:45.224199 kubelet[2683]: I1124 00:14:45.224194 2683 policy_none.go:49] "None policy: Start" Nov 24 00:14:45.224199 kubelet[2683]: I1124 00:14:45.224203 2683 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 24 00:14:45.224303 kubelet[2683]: I1124 00:14:45.224214 2683 state_mem.go:35] "Initializing new in-memory state store" Nov 24 00:14:45.224344 kubelet[2683]: I1124 00:14:45.224313 2683 state_mem.go:75] "Updated machine memory state" Nov 24 00:14:45.230037 kubelet[2683]: I1124 00:14:45.229980 2683 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 00:14:45.230212 kubelet[2683]: I1124 00:14:45.230191 2683 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 00:14:45.230254 kubelet[2683]: I1124 00:14:45.230203 2683 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 00:14:45.230453 kubelet[2683]: I1124 00:14:45.230428 2683 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 00:14:45.233887 kubelet[2683]: E1124 00:14:45.233841 2683 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 24 00:14:45.291104 kubelet[2683]: I1124 00:14:45.291035 2683 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:45.292198 kubelet[2683]: I1124 00:14:45.291689 2683 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:45.292198 kubelet[2683]: I1124 00:14:45.291873 2683 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Nov 24 00:14:45.342741 kubelet[2683]: I1124 00:14:45.342689 2683 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 24 00:14:45.379309 kubelet[2683]: E1124 00:14:45.378602 2683 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:45.475992 kubelet[2683]: I1124 00:14:45.475806 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/db6cd6cee75c80f34a5bc34c3c506b23-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"db6cd6cee75c80f34a5bc34c3c506b23\") " pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:45.475992 kubelet[2683]: I1124 00:14:45.475879 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:45.475992 kubelet[2683]: I1124 00:14:45.475908 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:45.475992 kubelet[2683]: I1124 00:14:45.475935 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0a68423804124305a9de061f38780871-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"0a68423804124305a9de061f38780871\") " pod="kube-system/kube-scheduler-localhost" Nov 24 00:14:45.475992 kubelet[2683]: I1124 00:14:45.475974 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/db6cd6cee75c80f34a5bc34c3c506b23-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"db6cd6cee75c80f34a5bc34c3c506b23\") " pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:45.476351 kubelet[2683]: I1124 00:14:45.475994 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/db6cd6cee75c80f34a5bc34c3c506b23-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"db6cd6cee75c80f34a5bc34c3c506b23\") " pod="kube-system/kube-apiserver-localhost" Nov 24 00:14:45.476351 kubelet[2683]: I1124 00:14:45.476016 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:45.476351 kubelet[2683]: I1124 00:14:45.476038 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:45.476351 kubelet[2683]: I1124 00:14:45.476059 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 24 00:14:45.516094 kubelet[2683]: I1124 00:14:45.516048 2683 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Nov 24 00:14:45.516244 kubelet[2683]: I1124 00:14:45.516159 2683 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Nov 24 00:14:45.664431 kubelet[2683]: E1124 00:14:45.664384 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:45.664431 kubelet[2683]: E1124 00:14:45.664384 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:45.679997 kubelet[2683]: E1124 00:14:45.679960 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:46.159580 kubelet[2683]: I1124 00:14:46.159510 2683 apiserver.go:52] "Watching apiserver" Nov 24 00:14:46.174951 kubelet[2683]: I1124 00:14:46.174894 2683 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 24 00:14:46.204970 kubelet[2683]: I1124 00:14:46.204911 2683 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Nov 24 00:14:46.205552 kubelet[2683]: E1124 00:14:46.205504 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:46.205989 kubelet[2683]: E1124 00:14:46.205955 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:46.211894 kubelet[2683]: E1124 00:14:46.211784 2683 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Nov 24 00:14:46.212116 kubelet[2683]: E1124 00:14:46.212004 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:46.236956 kubelet[2683]: I1124 00:14:46.236833 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.236810426 podStartE2EDuration="1.236810426s" podCreationTimestamp="2025-11-24 00:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:14:46.226919144 +0000 UTC m=+1.134564496" watchObservedRunningTime="2025-11-24 00:14:46.236810426 +0000 UTC m=+1.144455778" Nov 24 00:14:46.246726 kubelet[2683]: I1124 00:14:46.246036 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=3.246010131 podStartE2EDuration="3.246010131s" podCreationTimestamp="2025-11-24 00:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:14:46.245764573 +0000 UTC m=+1.153409925" watchObservedRunningTime="2025-11-24 00:14:46.246010131 +0000 UTC m=+1.153655483" Nov 24 00:14:46.246726 kubelet[2683]: I1124 00:14:46.246132 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.246127104 podStartE2EDuration="1.246127104s" podCreationTimestamp="2025-11-24 00:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:14:46.23678621 +0000 UTC m=+1.144431562" watchObservedRunningTime="2025-11-24 00:14:46.246127104 +0000 UTC m=+1.153772456" Nov 24 00:14:47.206610 kubelet[2683]: E1124 00:14:47.206569 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:47.207264 kubelet[2683]: E1124 00:14:47.206962 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:47.214667 kubelet[2683]: E1124 00:14:47.214597 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:48.209024 kubelet[2683]: E1124 00:14:48.208979 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:48.869989 kubelet[2683]: E1124 00:14:48.869930 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:50.251265 kubelet[2683]: I1124 00:14:50.251207 2683 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Nov 24 00:14:50.251944 containerd[1537]: time="2025-11-24T00:14:50.251856688Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Nov 24 00:14:50.252447 kubelet[2683]: I1124 00:14:50.252336 2683 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Nov 24 00:14:50.369185 update_engine[1522]: I20251124 00:14:50.369042 1522 update_attempter.cc:509] Updating boot flags... Nov 24 00:14:50.900202 systemd[1]: Created slice kubepods-besteffort-podc2434b93_00e8_43a1_9d0d_5ac6b62f4fbb.slice - libcontainer container kubepods-besteffort-podc2434b93_00e8_43a1_9d0d_5ac6b62f4fbb.slice. Nov 24 00:14:50.910521 kubelet[2683]: I1124 00:14:50.910469 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhdxw\" (UniqueName: \"kubernetes.io/projected/c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb-kube-api-access-fhdxw\") pod \"kube-proxy-wvzm9\" (UID: \"c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb\") " pod="kube-system/kube-proxy-wvzm9" Nov 24 00:14:50.910521 kubelet[2683]: I1124 00:14:50.910508 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb-kube-proxy\") pod \"kube-proxy-wvzm9\" (UID: \"c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb\") " pod="kube-system/kube-proxy-wvzm9" Nov 24 00:14:50.910745 kubelet[2683]: I1124 00:14:50.910548 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb-xtables-lock\") pod \"kube-proxy-wvzm9\" (UID: \"c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb\") " pod="kube-system/kube-proxy-wvzm9" Nov 24 00:14:50.910745 kubelet[2683]: I1124 00:14:50.910570 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb-lib-modules\") pod \"kube-proxy-wvzm9\" (UID: \"c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb\") " pod="kube-system/kube-proxy-wvzm9" Nov 24 00:14:51.018575 kubelet[2683]: E1124 00:14:51.018466 2683 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Nov 24 00:14:51.018575 kubelet[2683]: E1124 00:14:51.018503 2683 projected.go:194] Error preparing data for projected volume kube-api-access-fhdxw for pod kube-system/kube-proxy-wvzm9: configmap "kube-root-ca.crt" not found Nov 24 00:14:51.018575 kubelet[2683]: E1124 00:14:51.018570 2683 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb-kube-api-access-fhdxw podName:c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb nodeName:}" failed. No retries permitted until 2025-11-24 00:14:51.518548473 +0000 UTC m=+6.426193825 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-fhdxw" (UniqueName: "kubernetes.io/projected/c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb-kube-api-access-fhdxw") pod "kube-proxy-wvzm9" (UID: "c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb") : configmap "kube-root-ca.crt" not found Nov 24 00:14:51.301724 systemd[1]: Created slice kubepods-besteffort-pod0f99e40e_a974_4d63_89d1_8905219b806e.slice - libcontainer container kubepods-besteffort-pod0f99e40e_a974_4d63_89d1_8905219b806e.slice. Nov 24 00:14:51.313907 kubelet[2683]: I1124 00:14:51.313856 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szsjn\" (UniqueName: \"kubernetes.io/projected/0f99e40e-a974-4d63-89d1-8905219b806e-kube-api-access-szsjn\") pod \"tigera-operator-7dcd859c48-wjndq\" (UID: \"0f99e40e-a974-4d63-89d1-8905219b806e\") " pod="tigera-operator/tigera-operator-7dcd859c48-wjndq" Nov 24 00:14:51.313907 kubelet[2683]: I1124 00:14:51.313898 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/0f99e40e-a974-4d63-89d1-8905219b806e-var-lib-calico\") pod \"tigera-operator-7dcd859c48-wjndq\" (UID: \"0f99e40e-a974-4d63-89d1-8905219b806e\") " pod="tigera-operator/tigera-operator-7dcd859c48-wjndq" Nov 24 00:14:51.605888 containerd[1537]: time="2025-11-24T00:14:51.605767895Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-wjndq,Uid:0f99e40e-a974-4d63-89d1-8905219b806e,Namespace:tigera-operator,Attempt:0,}" Nov 24 00:14:51.811443 kubelet[2683]: E1124 00:14:51.811399 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:51.812013 containerd[1537]: time="2025-11-24T00:14:51.811948342Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wvzm9,Uid:c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb,Namespace:kube-system,Attempt:0,}" Nov 24 00:14:51.969659 containerd[1537]: time="2025-11-24T00:14:51.969426632Z" level=info msg="connecting to shim 51647411f6d266c530d1e5521e09d5a23af15a3912e97aecf71d1c004de87b88" address="unix:///run/containerd/s/b1054311375ed5c0381360cddd4dd92255adad1d9756d881a4b2e7b0f07df5bd" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:14:51.984574 containerd[1537]: time="2025-11-24T00:14:51.984051883Z" level=info msg="connecting to shim 55ab317a98f5d6e067d4253093e3e789c833eaa9cccfbf40ae58a0689638585b" address="unix:///run/containerd/s/d78faf6dd3792b341ee8f1778fde6d4be7056808617606c59c29ecc12df79613" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:14:52.015880 systemd[1]: Started cri-containerd-51647411f6d266c530d1e5521e09d5a23af15a3912e97aecf71d1c004de87b88.scope - libcontainer container 51647411f6d266c530d1e5521e09d5a23af15a3912e97aecf71d1c004de87b88. Nov 24 00:14:52.018257 systemd[1]: Started cri-containerd-55ab317a98f5d6e067d4253093e3e789c833eaa9cccfbf40ae58a0689638585b.scope - libcontainer container 55ab317a98f5d6e067d4253093e3e789c833eaa9cccfbf40ae58a0689638585b. Nov 24 00:14:52.106738 containerd[1537]: time="2025-11-24T00:14:52.106670611Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wvzm9,Uid:c2434b93-00e8-43a1-9d0d-5ac6b62f4fbb,Namespace:kube-system,Attempt:0,} returns sandbox id \"55ab317a98f5d6e067d4253093e3e789c833eaa9cccfbf40ae58a0689638585b\"" Nov 24 00:14:52.107225 kubelet[2683]: E1124 00:14:52.107194 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:52.109293 containerd[1537]: time="2025-11-24T00:14:52.109252285Z" level=info msg="CreateContainer within sandbox \"55ab317a98f5d6e067d4253093e3e789c833eaa9cccfbf40ae58a0689638585b\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Nov 24 00:14:52.188038 containerd[1537]: time="2025-11-24T00:14:52.187977914Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-wjndq,Uid:0f99e40e-a974-4d63-89d1-8905219b806e,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"51647411f6d266c530d1e5521e09d5a23af15a3912e97aecf71d1c004de87b88\"" Nov 24 00:14:52.189710 containerd[1537]: time="2025-11-24T00:14:52.189680692Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Nov 24 00:14:52.231169 containerd[1537]: time="2025-11-24T00:14:52.231030286Z" level=info msg="Container e6d72d563ef6a5e29d6cd285a6f8390de393ac40f69b0af684091070b420db8f: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:14:52.255745 containerd[1537]: time="2025-11-24T00:14:52.255643960Z" level=info msg="CreateContainer within sandbox \"55ab317a98f5d6e067d4253093e3e789c833eaa9cccfbf40ae58a0689638585b\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"e6d72d563ef6a5e29d6cd285a6f8390de393ac40f69b0af684091070b420db8f\"" Nov 24 00:14:52.256478 containerd[1537]: time="2025-11-24T00:14:52.256397699Z" level=info msg="StartContainer for \"e6d72d563ef6a5e29d6cd285a6f8390de393ac40f69b0af684091070b420db8f\"" Nov 24 00:14:52.258191 containerd[1537]: time="2025-11-24T00:14:52.258164397Z" level=info msg="connecting to shim e6d72d563ef6a5e29d6cd285a6f8390de393ac40f69b0af684091070b420db8f" address="unix:///run/containerd/s/d78faf6dd3792b341ee8f1778fde6d4be7056808617606c59c29ecc12df79613" protocol=ttrpc version=3 Nov 24 00:14:52.293889 systemd[1]: Started cri-containerd-e6d72d563ef6a5e29d6cd285a6f8390de393ac40f69b0af684091070b420db8f.scope - libcontainer container e6d72d563ef6a5e29d6cd285a6f8390de393ac40f69b0af684091070b420db8f. Nov 24 00:14:52.526232 containerd[1537]: time="2025-11-24T00:14:52.526174463Z" level=info msg="StartContainer for \"e6d72d563ef6a5e29d6cd285a6f8390de393ac40f69b0af684091070b420db8f\" returns successfully" Nov 24 00:14:53.220120 kubelet[2683]: E1124 00:14:53.220081 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:54.221927 kubelet[2683]: E1124 00:14:54.221893 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:56.399857 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3982867910.mount: Deactivated successfully. Nov 24 00:14:57.220412 kubelet[2683]: E1124 00:14:57.220358 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:57.232590 kubelet[2683]: I1124 00:14:57.232522 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-wvzm9" podStartSLOduration=7.232508108 podStartE2EDuration="7.232508108s" podCreationTimestamp="2025-11-24 00:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:14:53.283440722 +0000 UTC m=+8.191086074" watchObservedRunningTime="2025-11-24 00:14:57.232508108 +0000 UTC m=+12.140153460" Nov 24 00:14:57.289229 containerd[1537]: time="2025-11-24T00:14:57.289132252Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:57.290418 containerd[1537]: time="2025-11-24T00:14:57.290377195Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=25061691" Nov 24 00:14:57.291533 containerd[1537]: time="2025-11-24T00:14:57.291470612Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:57.294196 containerd[1537]: time="2025-11-24T00:14:57.294153081Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:14:57.295020 containerd[1537]: time="2025-11-24T00:14:57.294989401Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 5.10527673s" Nov 24 00:14:57.295062 containerd[1537]: time="2025-11-24T00:14:57.295021101Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Nov 24 00:14:57.296671 containerd[1537]: time="2025-11-24T00:14:57.296644780Z" level=info msg="CreateContainer within sandbox \"51647411f6d266c530d1e5521e09d5a23af15a3912e97aecf71d1c004de87b88\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Nov 24 00:14:57.306441 containerd[1537]: time="2025-11-24T00:14:57.306390849Z" level=info msg="Container 761d8df3d7430d7dafcd2a3cf607a3e3a71c2b5ef1d674b2aa66b7b8f4220f16: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:14:57.315174 containerd[1537]: time="2025-11-24T00:14:57.315121930Z" level=info msg="CreateContainer within sandbox \"51647411f6d266c530d1e5521e09d5a23af15a3912e97aecf71d1c004de87b88\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"761d8df3d7430d7dafcd2a3cf607a3e3a71c2b5ef1d674b2aa66b7b8f4220f16\"" Nov 24 00:14:57.315649 containerd[1537]: time="2025-11-24T00:14:57.315587159Z" level=info msg="StartContainer for \"761d8df3d7430d7dafcd2a3cf607a3e3a71c2b5ef1d674b2aa66b7b8f4220f16\"" Nov 24 00:14:57.316360 containerd[1537]: time="2025-11-24T00:14:57.316327729Z" level=info msg="connecting to shim 761d8df3d7430d7dafcd2a3cf607a3e3a71c2b5ef1d674b2aa66b7b8f4220f16" address="unix:///run/containerd/s/b1054311375ed5c0381360cddd4dd92255adad1d9756d881a4b2e7b0f07df5bd" protocol=ttrpc version=3 Nov 24 00:14:57.370824 systemd[1]: Started cri-containerd-761d8df3d7430d7dafcd2a3cf607a3e3a71c2b5ef1d674b2aa66b7b8f4220f16.scope - libcontainer container 761d8df3d7430d7dafcd2a3cf607a3e3a71c2b5ef1d674b2aa66b7b8f4220f16. Nov 24 00:14:57.403007 containerd[1537]: time="2025-11-24T00:14:57.402955116Z" level=info msg="StartContainer for \"761d8df3d7430d7dafcd2a3cf607a3e3a71c2b5ef1d674b2aa66b7b8f4220f16\" returns successfully" Nov 24 00:14:57.568296 kubelet[2683]: E1124 00:14:57.568257 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:59.185125 kubelet[2683]: E1124 00:14:59.185070 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:59.610050 kubelet[2683]: E1124 00:14:59.609759 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:14:59.610050 kubelet[2683]: I1124 00:14:59.609803 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-wjndq" podStartSLOduration=3.503373393 podStartE2EDuration="8.609787264s" podCreationTimestamp="2025-11-24 00:14:51 +0000 UTC" firstStartedPulling="2025-11-24 00:14:52.189246338 +0000 UTC m=+7.096891690" lastFinishedPulling="2025-11-24 00:14:57.295660209 +0000 UTC m=+12.203305561" observedRunningTime="2025-11-24 00:14:58.398200777 +0000 UTC m=+13.305846129" watchObservedRunningTime="2025-11-24 00:14:59.609787264 +0000 UTC m=+14.517432616" Nov 24 00:15:11.277421 sudo[1746]: pam_unix(sudo:session): session closed for user root Nov 24 00:15:11.279318 sshd[1745]: Connection closed by 10.0.0.1 port 42130 Nov 24 00:15:11.280258 sshd-session[1742]: pam_unix(sshd:session): session closed for user core Nov 24 00:15:11.285979 systemd[1]: sshd@6-10.0.0.25:22-10.0.0.1:42130.service: Deactivated successfully. Nov 24 00:15:11.288491 systemd[1]: session-7.scope: Deactivated successfully. Nov 24 00:15:11.288759 systemd[1]: session-7.scope: Consumed 5.275s CPU time, 231.1M memory peak. Nov 24 00:15:11.290121 systemd-logind[1519]: Session 7 logged out. Waiting for processes to exit. Nov 24 00:15:11.291812 systemd-logind[1519]: Removed session 7. Nov 24 00:15:14.108839 systemd[1]: Created slice kubepods-besteffort-podac2df468_7604_4f1c_8d3f_ac42fed21888.slice - libcontainer container kubepods-besteffort-podac2df468_7604_4f1c_8d3f_ac42fed21888.slice. Nov 24 00:15:14.167930 kubelet[2683]: I1124 00:15:14.167849 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m66zl\" (UniqueName: \"kubernetes.io/projected/ac2df468-7604-4f1c-8d3f-ac42fed21888-kube-api-access-m66zl\") pod \"calico-typha-6b78c6d55f-txxnm\" (UID: \"ac2df468-7604-4f1c-8d3f-ac42fed21888\") " pod="calico-system/calico-typha-6b78c6d55f-txxnm" Nov 24 00:15:14.167930 kubelet[2683]: I1124 00:15:14.167905 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac2df468-7604-4f1c-8d3f-ac42fed21888-tigera-ca-bundle\") pod \"calico-typha-6b78c6d55f-txxnm\" (UID: \"ac2df468-7604-4f1c-8d3f-ac42fed21888\") " pod="calico-system/calico-typha-6b78c6d55f-txxnm" Nov 24 00:15:14.167930 kubelet[2683]: I1124 00:15:14.167921 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/ac2df468-7604-4f1c-8d3f-ac42fed21888-typha-certs\") pod \"calico-typha-6b78c6d55f-txxnm\" (UID: \"ac2df468-7604-4f1c-8d3f-ac42fed21888\") " pod="calico-system/calico-typha-6b78c6d55f-txxnm" Nov 24 00:15:14.296478 systemd[1]: Created slice kubepods-besteffort-pod40a23094_dbec_4106_b833_fdb5db48bea9.slice - libcontainer container kubepods-besteffort-pod40a23094_dbec_4106_b833_fdb5db48bea9.slice. Nov 24 00:15:14.369665 kubelet[2683]: I1124 00:15:14.369241 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-lib-modules\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369665 kubelet[2683]: I1124 00:15:14.369284 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-267cj\" (UniqueName: \"kubernetes.io/projected/40a23094-dbec-4106-b833-fdb5db48bea9-kube-api-access-267cj\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369665 kubelet[2683]: I1124 00:15:14.369304 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/40a23094-dbec-4106-b833-fdb5db48bea9-node-certs\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369665 kubelet[2683]: I1124 00:15:14.369322 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-policysync\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369665 kubelet[2683]: I1124 00:15:14.369345 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-xtables-lock\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369955 kubelet[2683]: I1124 00:15:14.369366 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-cni-bin-dir\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369955 kubelet[2683]: I1124 00:15:14.369386 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-cni-net-dir\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369955 kubelet[2683]: I1124 00:15:14.369406 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40a23094-dbec-4106-b833-fdb5db48bea9-tigera-ca-bundle\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369955 kubelet[2683]: I1124 00:15:14.369429 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-cni-log-dir\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.369955 kubelet[2683]: I1124 00:15:14.369475 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-flexvol-driver-host\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.370138 kubelet[2683]: I1124 00:15:14.369514 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-var-lib-calico\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.370138 kubelet[2683]: I1124 00:15:14.369538 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/40a23094-dbec-4106-b833-fdb5db48bea9-var-run-calico\") pod \"calico-node-nh7wm\" (UID: \"40a23094-dbec-4106-b833-fdb5db48bea9\") " pod="calico-system/calico-node-nh7wm" Nov 24 00:15:14.414740 kubelet[2683]: E1124 00:15:14.414686 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:14.415924 containerd[1537]: time="2025-11-24T00:15:14.415877789Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6b78c6d55f-txxnm,Uid:ac2df468-7604-4f1c-8d3f-ac42fed21888,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:14.466280 containerd[1537]: time="2025-11-24T00:15:14.466199765Z" level=info msg="connecting to shim e38c5a6caa56e7ed5d576616539cf65be63983fc28747b0ba8f8ef46beabe407" address="unix:///run/containerd/s/1b600cca902116f1424c7d9268815d7b3ed2f9ffa80045f10c2b7b71c68541c4" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:14.481916 kubelet[2683]: E1124 00:15:14.481694 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.481916 kubelet[2683]: W1124 00:15:14.481730 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.481916 kubelet[2683]: E1124 00:15:14.481773 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.483235 kubelet[2683]: E1124 00:15:14.483210 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.483235 kubelet[2683]: W1124 00:15:14.483228 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.483483 kubelet[2683]: E1124 00:15:14.483240 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.496738 kubelet[2683]: E1124 00:15:14.496686 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.499661 kubelet[2683]: E1124 00:15:14.495595 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:14.499734 kubelet[2683]: W1124 00:15:14.496984 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.499734 kubelet[2683]: E1124 00:15:14.499726 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.501763 kubelet[2683]: E1124 00:15:14.501496 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.501763 kubelet[2683]: W1124 00:15:14.501518 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.501763 kubelet[2683]: E1124 00:15:14.501534 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.517973 systemd[1]: Started cri-containerd-e38c5a6caa56e7ed5d576616539cf65be63983fc28747b0ba8f8ef46beabe407.scope - libcontainer container e38c5a6caa56e7ed5d576616539cf65be63983fc28747b0ba8f8ef46beabe407. Nov 24 00:15:14.551245 kubelet[2683]: E1124 00:15:14.551123 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.551245 kubelet[2683]: W1124 00:15:14.551167 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.551245 kubelet[2683]: E1124 00:15:14.551197 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.551475 kubelet[2683]: E1124 00:15:14.551434 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.551475 kubelet[2683]: W1124 00:15:14.551443 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.551475 kubelet[2683]: E1124 00:15:14.551452 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.552499 kubelet[2683]: E1124 00:15:14.552452 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.552499 kubelet[2683]: W1124 00:15:14.552467 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.552499 kubelet[2683]: E1124 00:15:14.552478 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.552773 kubelet[2683]: E1124 00:15:14.552745 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.552773 kubelet[2683]: W1124 00:15:14.552760 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.552773 kubelet[2683]: E1124 00:15:14.552769 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.553648 kubelet[2683]: E1124 00:15:14.552999 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.553648 kubelet[2683]: W1124 00:15:14.553009 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.553648 kubelet[2683]: E1124 00:15:14.553018 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.553648 kubelet[2683]: E1124 00:15:14.553228 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.553648 kubelet[2683]: W1124 00:15:14.553237 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.553648 kubelet[2683]: E1124 00:15:14.553245 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.553648 kubelet[2683]: E1124 00:15:14.553429 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.553648 kubelet[2683]: W1124 00:15:14.553447 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.553648 kubelet[2683]: E1124 00:15:14.553455 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.554018 kubelet[2683]: E1124 00:15:14.553667 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.554018 kubelet[2683]: W1124 00:15:14.553675 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.554018 kubelet[2683]: E1124 00:15:14.553683 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.554018 kubelet[2683]: E1124 00:15:14.553890 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.554018 kubelet[2683]: W1124 00:15:14.553898 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.554018 kubelet[2683]: E1124 00:15:14.553906 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.554208 kubelet[2683]: E1124 00:15:14.554077 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.554208 kubelet[2683]: W1124 00:15:14.554084 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.554208 kubelet[2683]: E1124 00:15:14.554092 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.554836 kubelet[2683]: E1124 00:15:14.554317 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.554836 kubelet[2683]: W1124 00:15:14.554327 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.554836 kubelet[2683]: E1124 00:15:14.554335 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.554836 kubelet[2683]: E1124 00:15:14.554502 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.554836 kubelet[2683]: W1124 00:15:14.554509 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.554836 kubelet[2683]: E1124 00:15:14.554516 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.554836 kubelet[2683]: E1124 00:15:14.554770 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.554836 kubelet[2683]: W1124 00:15:14.554779 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.554836 kubelet[2683]: E1124 00:15:14.554788 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.555653 kubelet[2683]: E1124 00:15:14.555593 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.555653 kubelet[2683]: W1124 00:15:14.555607 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.555653 kubelet[2683]: E1124 00:15:14.555635 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.555972 kubelet[2683]: E1124 00:15:14.555954 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.555972 kubelet[2683]: W1124 00:15:14.555969 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.556061 kubelet[2683]: E1124 00:15:14.555979 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.556870 kubelet[2683]: E1124 00:15:14.556846 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.556870 kubelet[2683]: W1124 00:15:14.556858 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.556870 kubelet[2683]: E1124 00:15:14.556870 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.557122 kubelet[2683]: E1124 00:15:14.557096 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.557188 kubelet[2683]: W1124 00:15:14.557137 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.557188 kubelet[2683]: E1124 00:15:14.557155 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.557471 kubelet[2683]: E1124 00:15:14.557431 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.557525 kubelet[2683]: W1124 00:15:14.557486 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.557525 kubelet[2683]: E1124 00:15:14.557501 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.557863 kubelet[2683]: E1124 00:15:14.557839 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.557925 kubelet[2683]: W1124 00:15:14.557866 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.557925 kubelet[2683]: E1124 00:15:14.557878 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.558189 kubelet[2683]: E1124 00:15:14.558150 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.558189 kubelet[2683]: W1124 00:15:14.558176 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.558189 kubelet[2683]: E1124 00:15:14.558189 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.572535 kubelet[2683]: E1124 00:15:14.572481 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.572535 kubelet[2683]: W1124 00:15:14.572510 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.572535 kubelet[2683]: E1124 00:15:14.572544 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.572734 kubelet[2683]: I1124 00:15:14.572577 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7cdfd74-3050-4e07-b3f8-0110308ddcf6-kubelet-dir\") pod \"csi-node-driver-5jkmb\" (UID: \"e7cdfd74-3050-4e07-b3f8-0110308ddcf6\") " pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:14.573188 containerd[1537]: time="2025-11-24T00:15:14.573141347Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6b78c6d55f-txxnm,Uid:ac2df468-7604-4f1c-8d3f-ac42fed21888,Namespace:calico-system,Attempt:0,} returns sandbox id \"e38c5a6caa56e7ed5d576616539cf65be63983fc28747b0ba8f8ef46beabe407\"" Nov 24 00:15:14.573759 kubelet[2683]: E1124 00:15:14.573592 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.573759 kubelet[2683]: W1124 00:15:14.573604 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.573897 kubelet[2683]: E1124 00:15:14.573864 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.573897 kubelet[2683]: I1124 00:15:14.573889 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e7cdfd74-3050-4e07-b3f8-0110308ddcf6-registration-dir\") pod \"csi-node-driver-5jkmb\" (UID: \"e7cdfd74-3050-4e07-b3f8-0110308ddcf6\") " pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:14.573975 kubelet[2683]: E1124 00:15:14.573938 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.573975 kubelet[2683]: W1124 00:15:14.573951 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.573975 kubelet[2683]: E1124 00:15:14.573964 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.574224 kubelet[2683]: E1124 00:15:14.574184 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.574224 kubelet[2683]: W1124 00:15:14.574198 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.574224 kubelet[2683]: E1124 00:15:14.574210 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.574403 kubelet[2683]: E1124 00:15:14.574390 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.574403 kubelet[2683]: W1124 00:15:14.574400 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.574463 kubelet[2683]: E1124 00:15:14.574409 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.574463 kubelet[2683]: I1124 00:15:14.574424 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e7cdfd74-3050-4e07-b3f8-0110308ddcf6-socket-dir\") pod \"csi-node-driver-5jkmb\" (UID: \"e7cdfd74-3050-4e07-b3f8-0110308ddcf6\") " pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:14.574519 kubelet[2683]: E1124 00:15:14.574182 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:14.574670 kubelet[2683]: E1124 00:15:14.574651 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.574670 kubelet[2683]: W1124 00:15:14.574666 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.574865 kubelet[2683]: E1124 00:15:14.574675 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.574865 kubelet[2683]: I1124 00:15:14.574691 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/e7cdfd74-3050-4e07-b3f8-0110308ddcf6-varrun\") pod \"csi-node-driver-5jkmb\" (UID: \"e7cdfd74-3050-4e07-b3f8-0110308ddcf6\") " pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:14.575215 kubelet[2683]: E1124 00:15:14.575197 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.575215 kubelet[2683]: W1124 00:15:14.575211 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.575330 kubelet[2683]: E1124 00:15:14.575232 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.575493 kubelet[2683]: E1124 00:15:14.575444 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.575493 kubelet[2683]: W1124 00:15:14.575468 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.575493 kubelet[2683]: E1124 00:15:14.575481 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.575931 kubelet[2683]: E1124 00:15:14.575911 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.575931 kubelet[2683]: W1124 00:15:14.575924 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.576033 kubelet[2683]: E1124 00:15:14.575936 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.576033 kubelet[2683]: I1124 00:15:14.575957 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkmrt\" (UniqueName: \"kubernetes.io/projected/e7cdfd74-3050-4e07-b3f8-0110308ddcf6-kube-api-access-fkmrt\") pod \"csi-node-driver-5jkmb\" (UID: \"e7cdfd74-3050-4e07-b3f8-0110308ddcf6\") " pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:14.576221 kubelet[2683]: E1124 00:15:14.576160 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.576221 kubelet[2683]: W1124 00:15:14.576186 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.576363 containerd[1537]: time="2025-11-24T00:15:14.576277704Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Nov 24 00:15:14.576456 kubelet[2683]: E1124 00:15:14.576373 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.576456 kubelet[2683]: W1124 00:15:14.576382 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.576456 kubelet[2683]: E1124 00:15:14.576391 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.576456 kubelet[2683]: E1124 00:15:14.576429 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.576711 kubelet[2683]: E1124 00:15:14.576558 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.576711 kubelet[2683]: W1124 00:15:14.576566 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.576711 kubelet[2683]: E1124 00:15:14.576580 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.576983 kubelet[2683]: E1124 00:15:14.576768 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.576983 kubelet[2683]: W1124 00:15:14.576776 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.576983 kubelet[2683]: E1124 00:15:14.576786 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.576983 kubelet[2683]: E1124 00:15:14.576979 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.577372 kubelet[2683]: W1124 00:15:14.576989 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.577372 kubelet[2683]: E1124 00:15:14.576998 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.577372 kubelet[2683]: E1124 00:15:14.577180 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.577372 kubelet[2683]: W1124 00:15:14.577189 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.577372 kubelet[2683]: E1124 00:15:14.577199 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.599689 kubelet[2683]: E1124 00:15:14.599650 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:14.600299 containerd[1537]: time="2025-11-24T00:15:14.600258393Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-nh7wm,Uid:40a23094-dbec-4106-b833-fdb5db48bea9,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:14.677195 kubelet[2683]: E1124 00:15:14.677055 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.677195 kubelet[2683]: W1124 00:15:14.677081 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.677195 kubelet[2683]: E1124 00:15:14.677101 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.677872 kubelet[2683]: E1124 00:15:14.677852 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.677872 kubelet[2683]: W1124 00:15:14.677869 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.677994 kubelet[2683]: E1124 00:15:14.677897 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.678137 kubelet[2683]: E1124 00:15:14.678114 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.678137 kubelet[2683]: W1124 00:15:14.678130 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.678232 kubelet[2683]: E1124 00:15:14.678150 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.678384 kubelet[2683]: E1124 00:15:14.678369 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.678420 kubelet[2683]: W1124 00:15:14.678383 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.678420 kubelet[2683]: E1124 00:15:14.678402 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.678605 kubelet[2683]: E1124 00:15:14.678591 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.678605 kubelet[2683]: W1124 00:15:14.678604 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.678689 kubelet[2683]: E1124 00:15:14.678639 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.678975 kubelet[2683]: E1124 00:15:14.678957 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.678975 kubelet[2683]: W1124 00:15:14.678970 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.679157 kubelet[2683]: E1124 00:15:14.678984 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.679277 kubelet[2683]: E1124 00:15:14.679258 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.679277 kubelet[2683]: W1124 00:15:14.679273 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.679489 kubelet[2683]: E1124 00:15:14.679298 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.679538 kubelet[2683]: E1124 00:15:14.679521 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.679538 kubelet[2683]: W1124 00:15:14.679532 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.679619 kubelet[2683]: E1124 00:15:14.679550 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.679783 kubelet[2683]: E1124 00:15:14.679764 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.679783 kubelet[2683]: W1124 00:15:14.679777 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.679871 kubelet[2683]: E1124 00:15:14.679794 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.679993 kubelet[2683]: E1124 00:15:14.679974 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.679993 kubelet[2683]: W1124 00:15:14.679985 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.680156 kubelet[2683]: E1124 00:15:14.680000 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.680262 kubelet[2683]: E1124 00:15:14.680243 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.680262 kubelet[2683]: W1124 00:15:14.680257 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.680354 kubelet[2683]: E1124 00:15:14.680275 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.680481 kubelet[2683]: E1124 00:15:14.680460 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.680481 kubelet[2683]: W1124 00:15:14.680473 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.680567 kubelet[2683]: E1124 00:15:14.680489 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.680699 kubelet[2683]: E1124 00:15:14.680681 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.680699 kubelet[2683]: W1124 00:15:14.680693 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.680771 kubelet[2683]: E1124 00:15:14.680709 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.680908 kubelet[2683]: E1124 00:15:14.680889 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.680908 kubelet[2683]: W1124 00:15:14.680902 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.680992 kubelet[2683]: E1124 00:15:14.680918 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.681132 kubelet[2683]: E1124 00:15:14.681112 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.681132 kubelet[2683]: W1124 00:15:14.681124 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.681221 kubelet[2683]: E1124 00:15:14.681139 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.681345 kubelet[2683]: E1124 00:15:14.681326 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.681345 kubelet[2683]: W1124 00:15:14.681338 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.681424 kubelet[2683]: E1124 00:15:14.681353 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.681594 kubelet[2683]: E1124 00:15:14.681574 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.681594 kubelet[2683]: W1124 00:15:14.681586 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.681697 kubelet[2683]: E1124 00:15:14.681659 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.681823 kubelet[2683]: E1124 00:15:14.681804 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.681823 kubelet[2683]: W1124 00:15:14.681815 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.681917 kubelet[2683]: E1124 00:15:14.681831 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.682028 kubelet[2683]: E1124 00:15:14.682010 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.682028 kubelet[2683]: W1124 00:15:14.682021 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.682093 kubelet[2683]: E1124 00:15:14.682036 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.682270 kubelet[2683]: E1124 00:15:14.682241 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.682270 kubelet[2683]: W1124 00:15:14.682256 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.682344 kubelet[2683]: E1124 00:15:14.682271 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.682456 kubelet[2683]: E1124 00:15:14.682440 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.682456 kubelet[2683]: W1124 00:15:14.682452 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.682515 kubelet[2683]: E1124 00:15:14.682465 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.682693 kubelet[2683]: E1124 00:15:14.682674 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.682693 kubelet[2683]: W1124 00:15:14.682687 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.682774 kubelet[2683]: E1124 00:15:14.682702 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.682954 kubelet[2683]: E1124 00:15:14.682936 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.682954 kubelet[2683]: W1124 00:15:14.682949 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.683019 kubelet[2683]: E1124 00:15:14.682964 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.683211 kubelet[2683]: E1124 00:15:14.683191 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.683211 kubelet[2683]: W1124 00:15:14.683203 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.683300 kubelet[2683]: E1124 00:15:14.683220 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.683523 kubelet[2683]: E1124 00:15:14.683507 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.683523 kubelet[2683]: W1124 00:15:14.683520 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.683594 kubelet[2683]: E1124 00:15:14.683531 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:14.750351 kubelet[2683]: E1124 00:15:14.750315 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:14.750351 kubelet[2683]: W1124 00:15:14.750339 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:14.750351 kubelet[2683]: E1124 00:15:14.750363 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:15.172308 containerd[1537]: time="2025-11-24T00:15:15.172255536Z" level=info msg="connecting to shim 690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26" address="unix:///run/containerd/s/7f7049f7fa2eddffbd1af32ef80cfeea7699eafe3d373bcadc21d6e309b3d52d" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:15.199889 systemd[1]: Started cri-containerd-690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26.scope - libcontainer container 690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26. Nov 24 00:15:15.286460 containerd[1537]: time="2025-11-24T00:15:15.286411389Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-nh7wm,Uid:40a23094-dbec-4106-b833-fdb5db48bea9,Namespace:calico-system,Attempt:0,} returns sandbox id \"690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26\"" Nov 24 00:15:15.287014 kubelet[2683]: E1124 00:15:15.286989 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:16.189864 kubelet[2683]: E1124 00:15:16.189800 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:16.713715 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2774112977.mount: Deactivated successfully. Nov 24 00:15:17.724531 containerd[1537]: time="2025-11-24T00:15:17.724463457Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:17.725767 containerd[1537]: time="2025-11-24T00:15:17.725740708Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=35234628" Nov 24 00:15:17.727281 containerd[1537]: time="2025-11-24T00:15:17.727247459Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:17.730145 containerd[1537]: time="2025-11-24T00:15:17.730065415Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:17.730793 containerd[1537]: time="2025-11-24T00:15:17.730734743Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 3.154422123s" Nov 24 00:15:17.730793 containerd[1537]: time="2025-11-24T00:15:17.730777323Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Nov 24 00:15:17.732161 containerd[1537]: time="2025-11-24T00:15:17.732042280Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Nov 24 00:15:17.744585 containerd[1537]: time="2025-11-24T00:15:17.744522194Z" level=info msg="CreateContainer within sandbox \"e38c5a6caa56e7ed5d576616539cf65be63983fc28747b0ba8f8ef46beabe407\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Nov 24 00:15:17.755743 containerd[1537]: time="2025-11-24T00:15:17.755687979Z" level=info msg="Container f0c57014f8c148309dd7d75db46c667ea91f01dd34e0e16084399aa7b46de3ea: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:15:17.766644 containerd[1537]: time="2025-11-24T00:15:17.766574960Z" level=info msg="CreateContainer within sandbox \"e38c5a6caa56e7ed5d576616539cf65be63983fc28747b0ba8f8ef46beabe407\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"f0c57014f8c148309dd7d75db46c667ea91f01dd34e0e16084399aa7b46de3ea\"" Nov 24 00:15:17.767377 containerd[1537]: time="2025-11-24T00:15:17.767129432Z" level=info msg="StartContainer for \"f0c57014f8c148309dd7d75db46c667ea91f01dd34e0e16084399aa7b46de3ea\"" Nov 24 00:15:17.768509 containerd[1537]: time="2025-11-24T00:15:17.768478787Z" level=info msg="connecting to shim f0c57014f8c148309dd7d75db46c667ea91f01dd34e0e16084399aa7b46de3ea" address="unix:///run/containerd/s/1b600cca902116f1424c7d9268815d7b3ed2f9ffa80045f10c2b7b71c68541c4" protocol=ttrpc version=3 Nov 24 00:15:17.792793 systemd[1]: Started cri-containerd-f0c57014f8c148309dd7d75db46c667ea91f01dd34e0e16084399aa7b46de3ea.scope - libcontainer container f0c57014f8c148309dd7d75db46c667ea91f01dd34e0e16084399aa7b46de3ea. Nov 24 00:15:17.881806 containerd[1537]: time="2025-11-24T00:15:17.881746628Z" level=info msg="StartContainer for \"f0c57014f8c148309dd7d75db46c667ea91f01dd34e0e16084399aa7b46de3ea\" returns successfully" Nov 24 00:15:18.190238 kubelet[2683]: E1124 00:15:18.190185 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:18.286298 kubelet[2683]: E1124 00:15:18.286254 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:18.287496 kubelet[2683]: E1124 00:15:18.287453 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.287496 kubelet[2683]: W1124 00:15:18.287480 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.287594 kubelet[2683]: E1124 00:15:18.287502 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.287815 kubelet[2683]: E1124 00:15:18.287775 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.287815 kubelet[2683]: W1124 00:15:18.287797 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.287815 kubelet[2683]: E1124 00:15:18.287809 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.288332 kubelet[2683]: E1124 00:15:18.288068 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.288332 kubelet[2683]: W1124 00:15:18.288105 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.288332 kubelet[2683]: E1124 00:15:18.288126 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.288609 kubelet[2683]: E1124 00:15:18.288593 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.288730 kubelet[2683]: W1124 00:15:18.288709 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.288860 kubelet[2683]: E1124 00:15:18.288800 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.289220 kubelet[2683]: E1124 00:15:18.289178 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.289220 kubelet[2683]: W1124 00:15:18.289204 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.289220 kubelet[2683]: E1124 00:15:18.289216 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.289484 kubelet[2683]: E1124 00:15:18.289453 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.289484 kubelet[2683]: W1124 00:15:18.289464 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.289601 kubelet[2683]: E1124 00:15:18.289501 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.289786 kubelet[2683]: E1124 00:15:18.289766 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.289786 kubelet[2683]: W1124 00:15:18.289781 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.289897 kubelet[2683]: E1124 00:15:18.289795 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.290059 kubelet[2683]: E1124 00:15:18.290015 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.290059 kubelet[2683]: W1124 00:15:18.290027 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.290059 kubelet[2683]: E1124 00:15:18.290037 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.290281 kubelet[2683]: E1124 00:15:18.290264 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.290281 kubelet[2683]: W1124 00:15:18.290276 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.290384 kubelet[2683]: E1124 00:15:18.290287 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.290545 kubelet[2683]: E1124 00:15:18.290507 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.290545 kubelet[2683]: W1124 00:15:18.290521 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.290545 kubelet[2683]: E1124 00:15:18.290532 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.290794 kubelet[2683]: E1124 00:15:18.290777 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.290794 kubelet[2683]: W1124 00:15:18.290791 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.290845 kubelet[2683]: E1124 00:15:18.290802 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.291028 kubelet[2683]: E1124 00:15:18.291011 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.291028 kubelet[2683]: W1124 00:15:18.291025 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.291108 kubelet[2683]: E1124 00:15:18.291037 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.291286 kubelet[2683]: E1124 00:15:18.291258 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.291286 kubelet[2683]: W1124 00:15:18.291272 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.291286 kubelet[2683]: E1124 00:15:18.291282 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.291507 kubelet[2683]: E1124 00:15:18.291491 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.291507 kubelet[2683]: W1124 00:15:18.291507 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.291569 kubelet[2683]: E1124 00:15:18.291519 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.291767 kubelet[2683]: E1124 00:15:18.291744 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.291767 kubelet[2683]: W1124 00:15:18.291759 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.291873 kubelet[2683]: E1124 00:15:18.291771 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.296948 kubelet[2683]: I1124 00:15:18.296889 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6b78c6d55f-txxnm" podStartSLOduration=1.141166477 podStartE2EDuration="4.296874487s" podCreationTimestamp="2025-11-24 00:15:14 +0000 UTC" firstStartedPulling="2025-11-24 00:15:14.575908641 +0000 UTC m=+29.483553993" lastFinishedPulling="2025-11-24 00:15:17.731616631 +0000 UTC m=+32.639262003" observedRunningTime="2025-11-24 00:15:18.296516855 +0000 UTC m=+33.204162208" watchObservedRunningTime="2025-11-24 00:15:18.296874487 +0000 UTC m=+33.204519829" Nov 24 00:15:18.305001 kubelet[2683]: E1124 00:15:18.304968 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.305001 kubelet[2683]: W1124 00:15:18.304991 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.305159 kubelet[2683]: E1124 00:15:18.305012 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.305256 kubelet[2683]: E1124 00:15:18.305242 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.305256 kubelet[2683]: W1124 00:15:18.305254 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.305322 kubelet[2683]: E1124 00:15:18.305268 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.305651 kubelet[2683]: E1124 00:15:18.305595 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.305707 kubelet[2683]: W1124 00:15:18.305647 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.305707 kubelet[2683]: E1124 00:15:18.305680 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.305919 kubelet[2683]: E1124 00:15:18.305889 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.305919 kubelet[2683]: W1124 00:15:18.305905 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.305990 kubelet[2683]: E1124 00:15:18.305921 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.306370 kubelet[2683]: E1124 00:15:18.306350 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.306370 kubelet[2683]: W1124 00:15:18.306366 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.306459 kubelet[2683]: E1124 00:15:18.306385 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.306685 kubelet[2683]: E1124 00:15:18.306665 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.306685 kubelet[2683]: W1124 00:15:18.306680 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.306800 kubelet[2683]: E1124 00:15:18.306722 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.306928 kubelet[2683]: E1124 00:15:18.306892 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.306928 kubelet[2683]: W1124 00:15:18.306904 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.307000 kubelet[2683]: E1124 00:15:18.306941 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.307154 kubelet[2683]: E1124 00:15:18.307138 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.307154 kubelet[2683]: W1124 00:15:18.307150 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.307220 kubelet[2683]: E1124 00:15:18.307165 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.307387 kubelet[2683]: E1124 00:15:18.307372 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.307387 kubelet[2683]: W1124 00:15:18.307383 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.307470 kubelet[2683]: E1124 00:15:18.307400 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.307599 kubelet[2683]: E1124 00:15:18.307571 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.307599 kubelet[2683]: W1124 00:15:18.307587 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.307599 kubelet[2683]: E1124 00:15:18.307604 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.307901 kubelet[2683]: E1124 00:15:18.307887 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.307901 kubelet[2683]: W1124 00:15:18.307897 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.307968 kubelet[2683]: E1124 00:15:18.307911 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.308148 kubelet[2683]: E1124 00:15:18.308135 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.308148 kubelet[2683]: W1124 00:15:18.308145 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.308283 kubelet[2683]: E1124 00:15:18.308192 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.308334 kubelet[2683]: E1124 00:15:18.308321 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.308334 kubelet[2683]: W1124 00:15:18.308331 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.308404 kubelet[2683]: E1124 00:15:18.308381 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.308543 kubelet[2683]: E1124 00:15:18.308525 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.308573 kubelet[2683]: W1124 00:15:18.308541 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.308573 kubelet[2683]: E1124 00:15:18.308560 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.308794 kubelet[2683]: E1124 00:15:18.308777 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.308794 kubelet[2683]: W1124 00:15:18.308791 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.308874 kubelet[2683]: E1124 00:15:18.308808 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.309038 kubelet[2683]: E1124 00:15:18.309021 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.309067 kubelet[2683]: W1124 00:15:18.309036 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.309067 kubelet[2683]: E1124 00:15:18.309055 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.309327 kubelet[2683]: E1124 00:15:18.309312 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.309327 kubelet[2683]: W1124 00:15:18.309324 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.309389 kubelet[2683]: E1124 00:15:18.309336 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:18.309511 kubelet[2683]: E1124 00:15:18.309499 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:18.309511 kubelet[2683]: W1124 00:15:18.309507 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:18.309570 kubelet[2683]: E1124 00:15:18.309515 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.173340 containerd[1537]: time="2025-11-24T00:15:19.173267102Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:19.174203 containerd[1537]: time="2025-11-24T00:15:19.174137577Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=4446754" Nov 24 00:15:19.176003 containerd[1537]: time="2025-11-24T00:15:19.175940785Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:19.178181 containerd[1537]: time="2025-11-24T00:15:19.178119319Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:19.178897 containerd[1537]: time="2025-11-24T00:15:19.178845743Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 1.446710057s" Nov 24 00:15:19.178897 containerd[1537]: time="2025-11-24T00:15:19.178881761Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Nov 24 00:15:19.181138 containerd[1537]: time="2025-11-24T00:15:19.181104086Z" level=info msg="CreateContainer within sandbox \"690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Nov 24 00:15:19.196056 containerd[1537]: time="2025-11-24T00:15:19.194663593Z" level=info msg="Container 09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:15:19.208554 containerd[1537]: time="2025-11-24T00:15:19.208503206Z" level=info msg="CreateContainer within sandbox \"690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5\"" Nov 24 00:15:19.209228 containerd[1537]: time="2025-11-24T00:15:19.209190357Z" level=info msg="StartContainer for \"09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5\"" Nov 24 00:15:19.211318 containerd[1537]: time="2025-11-24T00:15:19.211287087Z" level=info msg="connecting to shim 09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5" address="unix:///run/containerd/s/7f7049f7fa2eddffbd1af32ef80cfeea7699eafe3d373bcadc21d6e309b3d52d" protocol=ttrpc version=3 Nov 24 00:15:19.246962 systemd[1]: Started cri-containerd-09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5.scope - libcontainer container 09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5. Nov 24 00:15:19.291396 kubelet[2683]: E1124 00:15:19.291340 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:19.297680 kubelet[2683]: E1124 00:15:19.297586 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.297680 kubelet[2683]: W1124 00:15:19.297606 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.297680 kubelet[2683]: E1124 00:15:19.297636 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.297946 kubelet[2683]: E1124 00:15:19.297848 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.297946 kubelet[2683]: W1124 00:15:19.297859 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.297946 kubelet[2683]: E1124 00:15:19.297870 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.298080 kubelet[2683]: E1124 00:15:19.298045 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.298080 kubelet[2683]: W1124 00:15:19.298058 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.298080 kubelet[2683]: E1124 00:15:19.298078 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.298296 kubelet[2683]: E1124 00:15:19.298263 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.298296 kubelet[2683]: W1124 00:15:19.298277 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.298296 kubelet[2683]: E1124 00:15:19.298285 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.298504 kubelet[2683]: E1124 00:15:19.298484 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.298504 kubelet[2683]: W1124 00:15:19.298496 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.298504 kubelet[2683]: E1124 00:15:19.298506 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.298705 kubelet[2683]: E1124 00:15:19.298686 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.298705 kubelet[2683]: W1124 00:15:19.298696 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.298705 kubelet[2683]: E1124 00:15:19.298704 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.298923 kubelet[2683]: E1124 00:15:19.298897 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.298964 kubelet[2683]: W1124 00:15:19.298915 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.298964 kubelet[2683]: E1124 00:15:19.298944 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.299218 kubelet[2683]: E1124 00:15:19.299184 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.299218 kubelet[2683]: W1124 00:15:19.299201 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.299218 kubelet[2683]: E1124 00:15:19.299235 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.299719 kubelet[2683]: E1124 00:15:19.299697 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.299719 kubelet[2683]: W1124 00:15:19.299714 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.299835 kubelet[2683]: E1124 00:15:19.299726 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.300198 kubelet[2683]: E1124 00:15:19.300170 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.300198 kubelet[2683]: W1124 00:15:19.300185 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.300198 kubelet[2683]: E1124 00:15:19.300197 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.300502 kubelet[2683]: E1124 00:15:19.300487 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.300502 kubelet[2683]: W1124 00:15:19.300500 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.300578 kubelet[2683]: E1124 00:15:19.300512 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.300783 kubelet[2683]: E1124 00:15:19.300753 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.300783 kubelet[2683]: W1124 00:15:19.300766 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.300783 kubelet[2683]: E1124 00:15:19.300777 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.301050 kubelet[2683]: E1124 00:15:19.301035 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.301050 kubelet[2683]: W1124 00:15:19.301048 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.301152 kubelet[2683]: E1124 00:15:19.301059 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.301313 kubelet[2683]: E1124 00:15:19.301299 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.301359 kubelet[2683]: W1124 00:15:19.301314 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.301359 kubelet[2683]: E1124 00:15:19.301325 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.302090 kubelet[2683]: E1124 00:15:19.302061 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.302090 kubelet[2683]: W1124 00:15:19.302088 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.302171 kubelet[2683]: E1124 00:15:19.302100 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.313535 kubelet[2683]: E1124 00:15:19.313477 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.313535 kubelet[2683]: W1124 00:15:19.313493 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.313535 kubelet[2683]: E1124 00:15:19.313508 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.313860 kubelet[2683]: E1124 00:15:19.313847 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.313860 kubelet[2683]: W1124 00:15:19.313859 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.313947 kubelet[2683]: E1124 00:15:19.313880 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.314171 kubelet[2683]: E1124 00:15:19.314157 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.314171 kubelet[2683]: W1124 00:15:19.314168 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.314238 kubelet[2683]: E1124 00:15:19.314188 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.314430 kubelet[2683]: E1124 00:15:19.314408 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.314462 kubelet[2683]: W1124 00:15:19.314429 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.314462 kubelet[2683]: E1124 00:15:19.314449 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.314679 kubelet[2683]: E1124 00:15:19.314665 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.314679 kubelet[2683]: W1124 00:15:19.314675 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.314740 kubelet[2683]: E1124 00:15:19.314687 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.314871 kubelet[2683]: E1124 00:15:19.314856 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.314908 kubelet[2683]: W1124 00:15:19.314877 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.314972 kubelet[2683]: E1124 00:15:19.314958 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.315124 kubelet[2683]: E1124 00:15:19.315105 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.315124 kubelet[2683]: W1124 00:15:19.315115 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.315196 kubelet[2683]: E1124 00:15:19.315182 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.315370 kubelet[2683]: E1124 00:15:19.315351 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.315370 kubelet[2683]: W1124 00:15:19.315361 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.315455 kubelet[2683]: E1124 00:15:19.315419 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.315603 kubelet[2683]: E1124 00:15:19.315588 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.315603 kubelet[2683]: W1124 00:15:19.315599 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.315678 kubelet[2683]: E1124 00:15:19.315610 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.315942 kubelet[2683]: E1124 00:15:19.315907 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.315985 kubelet[2683]: W1124 00:15:19.315940 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.315985 kubelet[2683]: E1124 00:15:19.315971 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.316222 kubelet[2683]: E1124 00:15:19.316204 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.316222 kubelet[2683]: W1124 00:15:19.316219 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.316279 kubelet[2683]: E1124 00:15:19.316239 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.316453 kubelet[2683]: E1124 00:15:19.316423 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.316453 kubelet[2683]: W1124 00:15:19.316435 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.316453 kubelet[2683]: E1124 00:15:19.316449 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.316835 kubelet[2683]: E1124 00:15:19.316776 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.316835 kubelet[2683]: W1124 00:15:19.316791 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.316835 kubelet[2683]: E1124 00:15:19.316809 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.317061 kubelet[2683]: E1124 00:15:19.317046 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.317111 kubelet[2683]: W1124 00:15:19.317059 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.317111 kubelet[2683]: E1124 00:15:19.317097 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.317404 kubelet[2683]: E1124 00:15:19.317380 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.317404 kubelet[2683]: W1124 00:15:19.317394 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.317495 kubelet[2683]: E1124 00:15:19.317410 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.317844 kubelet[2683]: E1124 00:15:19.317824 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.317844 kubelet[2683]: W1124 00:15:19.317842 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.317927 kubelet[2683]: E1124 00:15:19.317862 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.318201 kubelet[2683]: E1124 00:15:19.318184 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.318201 kubelet[2683]: W1124 00:15:19.318199 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.318283 kubelet[2683]: E1124 00:15:19.318218 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.318474 kubelet[2683]: E1124 00:15:19.318436 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 24 00:15:19.318474 kubelet[2683]: W1124 00:15:19.318452 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 24 00:15:19.318474 kubelet[2683]: E1124 00:15:19.318463 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 24 00:15:19.330030 containerd[1537]: time="2025-11-24T00:15:19.329972622Z" level=info msg="StartContainer for \"09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5\" returns successfully" Nov 24 00:15:19.344616 systemd[1]: cri-containerd-09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5.scope: Deactivated successfully. Nov 24 00:15:19.347282 containerd[1537]: time="2025-11-24T00:15:19.347241508Z" level=info msg="received container exit event container_id:\"09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5\" id:\"09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5\" pid:3394 exited_at:{seconds:1763943319 nanos:346682277}" Nov 24 00:15:19.384544 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-09a2b0d776b6f1638455bb11827e6345ae3c787157e6e53faaf53eea8f4775c5-rootfs.mount: Deactivated successfully. Nov 24 00:15:20.190736 kubelet[2683]: E1124 00:15:20.189881 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:20.300555 kubelet[2683]: E1124 00:15:20.300513 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:20.301017 kubelet[2683]: E1124 00:15:20.300710 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:20.301707 containerd[1537]: time="2025-11-24T00:15:20.301554935Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Nov 24 00:15:22.190025 kubelet[2683]: E1124 00:15:22.189965 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:24.190209 kubelet[2683]: E1124 00:15:24.190131 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:25.818138 containerd[1537]: time="2025-11-24T00:15:25.818063749Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:25.819498 containerd[1537]: time="2025-11-24T00:15:25.819457526Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70446859" Nov 24 00:15:25.820759 containerd[1537]: time="2025-11-24T00:15:25.820721610Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:25.823061 containerd[1537]: time="2025-11-24T00:15:25.823013213Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:25.823597 containerd[1537]: time="2025-11-24T00:15:25.823561873Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 5.521959189s" Nov 24 00:15:25.823660 containerd[1537]: time="2025-11-24T00:15:25.823598161Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Nov 24 00:15:25.825981 containerd[1537]: time="2025-11-24T00:15:25.825885116Z" level=info msg="CreateContainer within sandbox \"690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Nov 24 00:15:25.834979 containerd[1537]: time="2025-11-24T00:15:25.834905473Z" level=info msg="Container 39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:15:25.845691 containerd[1537]: time="2025-11-24T00:15:25.845657293Z" level=info msg="CreateContainer within sandbox \"690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af\"" Nov 24 00:15:25.846250 containerd[1537]: time="2025-11-24T00:15:25.846185574Z" level=info msg="StartContainer for \"39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af\"" Nov 24 00:15:25.848089 containerd[1537]: time="2025-11-24T00:15:25.848053142Z" level=info msg="connecting to shim 39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af" address="unix:///run/containerd/s/7f7049f7fa2eddffbd1af32ef80cfeea7699eafe3d373bcadc21d6e309b3d52d" protocol=ttrpc version=3 Nov 24 00:15:25.874103 systemd[1]: Started cri-containerd-39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af.scope - libcontainer container 39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af. Nov 24 00:15:26.140803 containerd[1537]: time="2025-11-24T00:15:26.140615728Z" level=info msg="StartContainer for \"39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af\" returns successfully" Nov 24 00:15:26.190781 kubelet[2683]: E1124 00:15:26.190725 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:26.314100 kubelet[2683]: E1124 00:15:26.314065 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:27.315924 kubelet[2683]: E1124 00:15:27.315883 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:28.190092 kubelet[2683]: E1124 00:15:28.190042 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:28.217134 containerd[1537]: time="2025-11-24T00:15:28.217075660Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 24 00:15:28.219874 systemd[1]: cri-containerd-39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af.scope: Deactivated successfully. Nov 24 00:15:28.220205 systemd[1]: cri-containerd-39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af.scope: Consumed 653ms CPU time, 178.1M memory peak, 3M read from disk, 171.3M written to disk. Nov 24 00:15:28.222192 containerd[1537]: time="2025-11-24T00:15:28.222156128Z" level=info msg="received container exit event container_id:\"39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af\" id:\"39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af\" pid:3489 exited_at:{seconds:1763943328 nanos:221938510}" Nov 24 00:15:28.245867 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-39994a2b148d47405399bf499aa48932f5c097da5bf4799b1a7c4255c7d620af-rootfs.mount: Deactivated successfully. Nov 24 00:15:28.287134 kubelet[2683]: I1124 00:15:28.287096 2683 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Nov 24 00:15:29.158933 systemd[1]: Created slice kubepods-burstable-podd27f4aa2_b9ea_448e_a394_15df266e8761.slice - libcontainer container kubepods-burstable-podd27f4aa2_b9ea_448e_a394_15df266e8761.slice. Nov 24 00:15:29.169912 systemd[1]: Created slice kubepods-besteffort-pod3b61fcb5_22ca_4d05_817b_ade781b1a990.slice - libcontainer container kubepods-besteffort-pod3b61fcb5_22ca_4d05_817b_ade781b1a990.slice. Nov 24 00:15:29.177384 systemd[1]: Created slice kubepods-besteffort-pod6cfea110_fbd7_44cd_904d_4c7d268ca542.slice - libcontainer container kubepods-besteffort-pod6cfea110_fbd7_44cd_904d_4c7d268ca542.slice. Nov 24 00:15:29.184137 systemd[1]: Created slice kubepods-besteffort-pod024a87f6_fa17_474a_b3f9_39d35fa916e3.slice - libcontainer container kubepods-besteffort-pod024a87f6_fa17_474a_b3f9_39d35fa916e3.slice. Nov 24 00:15:29.190996 systemd[1]: Created slice kubepods-burstable-pod6978f440_5fc4_460d_8ab7_a825799d091f.slice - libcontainer container kubepods-burstable-pod6978f440_5fc4_460d_8ab7_a825799d091f.slice. Nov 24 00:15:29.196291 systemd[1]: Created slice kubepods-besteffort-pod22c38642_d71a_4cc4_996b_dcdaded5d62c.slice - libcontainer container kubepods-besteffort-pod22c38642_d71a_4cc4_996b_dcdaded5d62c.slice. Nov 24 00:15:29.201325 systemd[1]: Created slice kubepods-besteffort-podbbc561c8_9195_483f_8404_7885126b6cd1.slice - libcontainer container kubepods-besteffort-podbbc561c8_9195_483f_8404_7885126b6cd1.slice. Nov 24 00:15:29.249546 kubelet[2683]: I1124 00:15:29.249495 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3b61fcb5-22ca-4d05-817b-ade781b1a990-calico-apiserver-certs\") pod \"calico-apiserver-c779694df-9w4qh\" (UID: \"3b61fcb5-22ca-4d05-817b-ade781b1a990\") " pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" Nov 24 00:15:29.249546 kubelet[2683]: I1124 00:15:29.249542 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbpm4\" (UniqueName: \"kubernetes.io/projected/3b61fcb5-22ca-4d05-817b-ade781b1a990-kube-api-access-rbpm4\") pod \"calico-apiserver-c779694df-9w4qh\" (UID: \"3b61fcb5-22ca-4d05-817b-ade781b1a990\") " pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" Nov 24 00:15:29.250092 kubelet[2683]: I1124 00:15:29.249563 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/024a87f6-fa17-474a-b3f9-39d35fa916e3-tigera-ca-bundle\") pod \"calico-kube-controllers-5889ff9675-8zv2t\" (UID: \"024a87f6-fa17-474a-b3f9-39d35fa916e3\") " pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" Nov 24 00:15:29.250092 kubelet[2683]: I1124 00:15:29.249585 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn2qq\" (UniqueName: \"kubernetes.io/projected/024a87f6-fa17-474a-b3f9-39d35fa916e3-kube-api-access-xn2qq\") pod \"calico-kube-controllers-5889ff9675-8zv2t\" (UID: \"024a87f6-fa17-474a-b3f9-39d35fa916e3\") " pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" Nov 24 00:15:29.250092 kubelet[2683]: I1124 00:15:29.249641 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/22c38642-d71a-4cc4-996b-dcdaded5d62c-goldmane-key-pair\") pod \"goldmane-666569f655-hr2wl\" (UID: \"22c38642-d71a-4cc4-996b-dcdaded5d62c\") " pod="calico-system/goldmane-666569f655-hr2wl" Nov 24 00:15:29.250092 kubelet[2683]: I1124 00:15:29.249748 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vllsh\" (UniqueName: \"kubernetes.io/projected/bbc561c8-9195-483f-8404-7885126b6cd1-kube-api-access-vllsh\") pod \"whisker-69d76575d7-wx7rs\" (UID: \"bbc561c8-9195-483f-8404-7885126b6cd1\") " pod="calico-system/whisker-69d76575d7-wx7rs" Nov 24 00:15:29.250092 kubelet[2683]: I1124 00:15:29.249814 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/6cfea110-fbd7-44cd-904d-4c7d268ca542-calico-apiserver-certs\") pod \"calico-apiserver-c779694df-n2448\" (UID: \"6cfea110-fbd7-44cd-904d-4c7d268ca542\") " pod="calico-apiserver/calico-apiserver-c779694df-n2448" Nov 24 00:15:29.250224 kubelet[2683]: I1124 00:15:29.249869 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t59pv\" (UniqueName: \"kubernetes.io/projected/22c38642-d71a-4cc4-996b-dcdaded5d62c-kube-api-access-t59pv\") pod \"goldmane-666569f655-hr2wl\" (UID: \"22c38642-d71a-4cc4-996b-dcdaded5d62c\") " pod="calico-system/goldmane-666569f655-hr2wl" Nov 24 00:15:29.250224 kubelet[2683]: I1124 00:15:29.249888 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d27f4aa2-b9ea-448e-a394-15df266e8761-config-volume\") pod \"coredns-668d6bf9bc-fskrk\" (UID: \"d27f4aa2-b9ea-448e-a394-15df266e8761\") " pod="kube-system/coredns-668d6bf9bc-fskrk" Nov 24 00:15:29.250224 kubelet[2683]: I1124 00:15:29.249910 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/bbc561c8-9195-483f-8404-7885126b6cd1-whisker-backend-key-pair\") pod \"whisker-69d76575d7-wx7rs\" (UID: \"bbc561c8-9195-483f-8404-7885126b6cd1\") " pod="calico-system/whisker-69d76575d7-wx7rs" Nov 24 00:15:29.250224 kubelet[2683]: I1124 00:15:29.249925 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbc561c8-9195-483f-8404-7885126b6cd1-whisker-ca-bundle\") pod \"whisker-69d76575d7-wx7rs\" (UID: \"bbc561c8-9195-483f-8404-7885126b6cd1\") " pod="calico-system/whisker-69d76575d7-wx7rs" Nov 24 00:15:29.250224 kubelet[2683]: I1124 00:15:29.249939 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22c38642-d71a-4cc4-996b-dcdaded5d62c-goldmane-ca-bundle\") pod \"goldmane-666569f655-hr2wl\" (UID: \"22c38642-d71a-4cc4-996b-dcdaded5d62c\") " pod="calico-system/goldmane-666569f655-hr2wl" Nov 24 00:15:29.250355 kubelet[2683]: I1124 00:15:29.249987 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6978f440-5fc4-460d-8ab7-a825799d091f-config-volume\") pod \"coredns-668d6bf9bc-csrvc\" (UID: \"6978f440-5fc4-460d-8ab7-a825799d091f\") " pod="kube-system/coredns-668d6bf9bc-csrvc" Nov 24 00:15:29.250355 kubelet[2683]: I1124 00:15:29.250021 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbz2f\" (UniqueName: \"kubernetes.io/projected/6cfea110-fbd7-44cd-904d-4c7d268ca542-kube-api-access-hbz2f\") pod \"calico-apiserver-c779694df-n2448\" (UID: \"6cfea110-fbd7-44cd-904d-4c7d268ca542\") " pod="calico-apiserver/calico-apiserver-c779694df-n2448" Nov 24 00:15:29.250355 kubelet[2683]: I1124 00:15:29.250051 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c38642-d71a-4cc4-996b-dcdaded5d62c-config\") pod \"goldmane-666569f655-hr2wl\" (UID: \"22c38642-d71a-4cc4-996b-dcdaded5d62c\") " pod="calico-system/goldmane-666569f655-hr2wl" Nov 24 00:15:29.250355 kubelet[2683]: I1124 00:15:29.250106 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtjss\" (UniqueName: \"kubernetes.io/projected/6978f440-5fc4-460d-8ab7-a825799d091f-kube-api-access-jtjss\") pod \"coredns-668d6bf9bc-csrvc\" (UID: \"6978f440-5fc4-460d-8ab7-a825799d091f\") " pod="kube-system/coredns-668d6bf9bc-csrvc" Nov 24 00:15:29.250355 kubelet[2683]: I1124 00:15:29.250139 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np9nh\" (UniqueName: \"kubernetes.io/projected/d27f4aa2-b9ea-448e-a394-15df266e8761-kube-api-access-np9nh\") pod \"coredns-668d6bf9bc-fskrk\" (UID: \"d27f4aa2-b9ea-448e-a394-15df266e8761\") " pod="kube-system/coredns-668d6bf9bc-fskrk" Nov 24 00:15:29.326015 kubelet[2683]: E1124 00:15:29.325957 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:29.327211 containerd[1537]: time="2025-11-24T00:15:29.327117428Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Nov 24 00:15:29.464867 kubelet[2683]: E1124 00:15:29.464730 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:29.465544 containerd[1537]: time="2025-11-24T00:15:29.465403177Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fskrk,Uid:d27f4aa2-b9ea-448e-a394-15df266e8761,Namespace:kube-system,Attempt:0,}" Nov 24 00:15:29.474607 containerd[1537]: time="2025-11-24T00:15:29.474569746Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-9w4qh,Uid:3b61fcb5-22ca-4d05-817b-ade781b1a990,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:15:29.481103 containerd[1537]: time="2025-11-24T00:15:29.481036554Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-n2448,Uid:6cfea110-fbd7-44cd-904d-4c7d268ca542,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:15:29.488612 containerd[1537]: time="2025-11-24T00:15:29.488582009Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5889ff9675-8zv2t,Uid:024a87f6-fa17-474a-b3f9-39d35fa916e3,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:29.494944 kubelet[2683]: E1124 00:15:29.494874 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:29.495217 containerd[1537]: time="2025-11-24T00:15:29.495181938Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-csrvc,Uid:6978f440-5fc4-460d-8ab7-a825799d091f,Namespace:kube-system,Attempt:0,}" Nov 24 00:15:29.499045 containerd[1537]: time="2025-11-24T00:15:29.498988705Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-hr2wl,Uid:22c38642-d71a-4cc4-996b-dcdaded5d62c,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:29.505365 containerd[1537]: time="2025-11-24T00:15:29.505296976Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-69d76575d7-wx7rs,Uid:bbc561c8-9195-483f-8404-7885126b6cd1,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:29.824892 containerd[1537]: time="2025-11-24T00:15:29.824665966Z" level=error msg="Failed to destroy network for sandbox \"0507962c7c730966c4c87d74b47e1a9479435570747bc1b435d2cbf253725083\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.844050 containerd[1537]: time="2025-11-24T00:15:29.843849528Z" level=error msg="Failed to destroy network for sandbox \"b39e1567613c6ba382e058a47879c33a6a7a49fd62f17a05a22aeffb2f5f709d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.845004 containerd[1537]: time="2025-11-24T00:15:29.844962397Z" level=error msg="Failed to destroy network for sandbox \"074e76883f6087b2b6644763b28c1044d52b5d23b9208e5a3edb92da212ab6d3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.845204 containerd[1537]: time="2025-11-24T00:15:29.845120073Z" level=error msg="Failed to destroy network for sandbox \"bb02396ab06e21aac0cf170cded3a4e4ec2a50f36055989e73447962c1b1eb60\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.872334 containerd[1537]: time="2025-11-24T00:15:29.859944360Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fskrk,Uid:d27f4aa2-b9ea-448e-a394-15df266e8761,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bb02396ab06e21aac0cf170cded3a4e4ec2a50f36055989e73447962c1b1eb60\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.872534 containerd[1537]: time="2025-11-24T00:15:29.859981981Z" level=error msg="Failed to destroy network for sandbox \"c22a00fa868d9a663a4a154efa22a938bb67dc53838412587f20e4581053b867\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.872699 containerd[1537]: time="2025-11-24T00:15:29.860009313Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-n2448,Uid:6cfea110-fbd7-44cd-904d-4c7d268ca542,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0507962c7c730966c4c87d74b47e1a9479435570747bc1b435d2cbf253725083\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.872789 containerd[1537]: time="2025-11-24T00:15:29.860060750Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-9w4qh,Uid:3b61fcb5-22ca-4d05-817b-ade781b1a990,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"074e76883f6087b2b6644763b28c1044d52b5d23b9208e5a3edb92da212ab6d3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.872789 containerd[1537]: time="2025-11-24T00:15:29.860100665Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-69d76575d7-wx7rs,Uid:bbc561c8-9195-483f-8404-7885126b6cd1,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b39e1567613c6ba382e058a47879c33a6a7a49fd62f17a05a22aeffb2f5f709d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.872946 containerd[1537]: time="2025-11-24T00:15:29.860269321Z" level=error msg="Failed to destroy network for sandbox \"3cc16d87bc4aa390e6d03a20d50db7feebc048b322c0ca248dc6778b1a6895bf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.872946 containerd[1537]: time="2025-11-24T00:15:29.863378297Z" level=error msg="Failed to destroy network for sandbox \"6a5e13f2df64930957b13a4b2bb8dcca26cc3347f7fd1ecae73a3a55cf83cb0c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.874639 containerd[1537]: time="2025-11-24T00:15:29.874584182Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-hr2wl,Uid:22c38642-d71a-4cc4-996b-dcdaded5d62c,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c22a00fa868d9a663a4a154efa22a938bb67dc53838412587f20e4581053b867\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.876014 containerd[1537]: time="2025-11-24T00:15:29.875910893Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-csrvc,Uid:6978f440-5fc4-460d-8ab7-a825799d091f,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a5e13f2df64930957b13a4b2bb8dcca26cc3347f7fd1ecae73a3a55cf83cb0c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.877432 containerd[1537]: time="2025-11-24T00:15:29.877385001Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5889ff9675-8zv2t,Uid:024a87f6-fa17-474a-b3f9-39d35fa916e3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cc16d87bc4aa390e6d03a20d50db7feebc048b322c0ca248dc6778b1a6895bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.878995 kubelet[2683]: E1124 00:15:29.878656 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bb02396ab06e21aac0cf170cded3a4e4ec2a50f36055989e73447962c1b1eb60\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.878995 kubelet[2683]: E1124 00:15:29.878701 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a5e13f2df64930957b13a4b2bb8dcca26cc3347f7fd1ecae73a3a55cf83cb0c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.878995 kubelet[2683]: E1124 00:15:29.878659 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"074e76883f6087b2b6644763b28c1044d52b5d23b9208e5a3edb92da212ab6d3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.878995 kubelet[2683]: E1124 00:15:29.878717 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0507962c7c730966c4c87d74b47e1a9479435570747bc1b435d2cbf253725083\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.879313 kubelet[2683]: E1124 00:15:29.878743 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"074e76883f6087b2b6644763b28c1044d52b5d23b9208e5a3edb92da212ab6d3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" Nov 24 00:15:29.879313 kubelet[2683]: E1124 00:15:29.878744 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bb02396ab06e21aac0cf170cded3a4e4ec2a50f36055989e73447962c1b1eb60\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-fskrk" Nov 24 00:15:29.879313 kubelet[2683]: E1124 00:15:29.878754 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0507962c7c730966c4c87d74b47e1a9479435570747bc1b435d2cbf253725083\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c779694df-n2448" Nov 24 00:15:29.879313 kubelet[2683]: E1124 00:15:29.878766 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bb02396ab06e21aac0cf170cded3a4e4ec2a50f36055989e73447962c1b1eb60\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-fskrk" Nov 24 00:15:29.879459 kubelet[2683]: E1124 00:15:29.878781 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0507962c7c730966c4c87d74b47e1a9479435570747bc1b435d2cbf253725083\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c779694df-n2448" Nov 24 00:15:29.879459 kubelet[2683]: E1124 00:15:29.878822 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a5e13f2df64930957b13a4b2bb8dcca26cc3347f7fd1ecae73a3a55cf83cb0c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-csrvc" Nov 24 00:15:29.879459 kubelet[2683]: E1124 00:15:29.878656 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cc16d87bc4aa390e6d03a20d50db7feebc048b322c0ca248dc6778b1a6895bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.879459 kubelet[2683]: E1124 00:15:29.878679 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c22a00fa868d9a663a4a154efa22a938bb67dc53838412587f20e4581053b867\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.879590 kubelet[2683]: E1124 00:15:29.878848 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a5e13f2df64930957b13a4b2bb8dcca26cc3347f7fd1ecae73a3a55cf83cb0c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-csrvc" Nov 24 00:15:29.879590 kubelet[2683]: E1124 00:15:29.878859 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cc16d87bc4aa390e6d03a20d50db7feebc048b322c0ca248dc6778b1a6895bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" Nov 24 00:15:29.879590 kubelet[2683]: E1124 00:15:29.878861 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c22a00fa868d9a663a4a154efa22a938bb67dc53838412587f20e4581053b867\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-hr2wl" Nov 24 00:15:29.879590 kubelet[2683]: E1124 00:15:29.878870 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cc16d87bc4aa390e6d03a20d50db7feebc048b322c0ca248dc6778b1a6895bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" Nov 24 00:15:29.879754 kubelet[2683]: E1124 00:15:29.878876 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c22a00fa868d9a663a4a154efa22a938bb67dc53838412587f20e4581053b867\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-hr2wl" Nov 24 00:15:29.879754 kubelet[2683]: E1124 00:15:29.878881 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-csrvc_kube-system(6978f440-5fc4-460d-8ab7-a825799d091f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-csrvc_kube-system(6978f440-5fc4-460d-8ab7-a825799d091f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6a5e13f2df64930957b13a4b2bb8dcca26cc3347f7fd1ecae73a3a55cf83cb0c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-csrvc" podUID="6978f440-5fc4-460d-8ab7-a825799d091f" Nov 24 00:15:29.879754 kubelet[2683]: E1124 00:15:29.878896 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-hr2wl_calico-system(22c38642-d71a-4cc4-996b-dcdaded5d62c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-hr2wl_calico-system(22c38642-d71a-4cc4-996b-dcdaded5d62c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c22a00fa868d9a663a4a154efa22a938bb67dc53838412587f20e4581053b867\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-hr2wl" podUID="22c38642-d71a-4cc4-996b-dcdaded5d62c" Nov 24 00:15:29.879923 kubelet[2683]: E1124 00:15:29.878764 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"074e76883f6087b2b6644763b28c1044d52b5d23b9208e5a3edb92da212ab6d3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" Nov 24 00:15:29.879923 kubelet[2683]: E1124 00:15:29.878895 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5889ff9675-8zv2t_calico-system(024a87f6-fa17-474a-b3f9-39d35fa916e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5889ff9675-8zv2t_calico-system(024a87f6-fa17-474a-b3f9-39d35fa916e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3cc16d87bc4aa390e6d03a20d50db7feebc048b322c0ca248dc6778b1a6895bf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" podUID="024a87f6-fa17-474a-b3f9-39d35fa916e3" Nov 24 00:15:29.879923 kubelet[2683]: E1124 00:15:29.878828 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-fskrk_kube-system(d27f4aa2-b9ea-448e-a394-15df266e8761)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-fskrk_kube-system(d27f4aa2-b9ea-448e-a394-15df266e8761)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bb02396ab06e21aac0cf170cded3a4e4ec2a50f36055989e73447962c1b1eb60\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-fskrk" podUID="d27f4aa2-b9ea-448e-a394-15df266e8761" Nov 24 00:15:29.880068 kubelet[2683]: E1124 00:15:29.878704 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b39e1567613c6ba382e058a47879c33a6a7a49fd62f17a05a22aeffb2f5f709d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:29.880068 kubelet[2683]: E1124 00:15:29.878922 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c779694df-9w4qh_calico-apiserver(3b61fcb5-22ca-4d05-817b-ade781b1a990)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c779694df-9w4qh_calico-apiserver(3b61fcb5-22ca-4d05-817b-ade781b1a990)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"074e76883f6087b2b6644763b28c1044d52b5d23b9208e5a3edb92da212ab6d3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" podUID="3b61fcb5-22ca-4d05-817b-ade781b1a990" Nov 24 00:15:29.880068 kubelet[2683]: E1124 00:15:29.878929 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b39e1567613c6ba382e058a47879c33a6a7a49fd62f17a05a22aeffb2f5f709d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-69d76575d7-wx7rs" Nov 24 00:15:29.880201 kubelet[2683]: E1124 00:15:29.878942 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b39e1567613c6ba382e058a47879c33a6a7a49fd62f17a05a22aeffb2f5f709d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-69d76575d7-wx7rs" Nov 24 00:15:29.880201 kubelet[2683]: E1124 00:15:29.878961 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-69d76575d7-wx7rs_calico-system(bbc561c8-9195-483f-8404-7885126b6cd1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-69d76575d7-wx7rs_calico-system(bbc561c8-9195-483f-8404-7885126b6cd1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b39e1567613c6ba382e058a47879c33a6a7a49fd62f17a05a22aeffb2f5f709d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-69d76575d7-wx7rs" podUID="bbc561c8-9195-483f-8404-7885126b6cd1" Nov 24 00:15:29.880201 kubelet[2683]: E1124 00:15:29.878828 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c779694df-n2448_calico-apiserver(6cfea110-fbd7-44cd-904d-4c7d268ca542)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c779694df-n2448_calico-apiserver(6cfea110-fbd7-44cd-904d-4c7d268ca542)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0507962c7c730966c4c87d74b47e1a9479435570747bc1b435d2cbf253725083\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c779694df-n2448" podUID="6cfea110-fbd7-44cd-904d-4c7d268ca542" Nov 24 00:15:30.196274 systemd[1]: Created slice kubepods-besteffort-pode7cdfd74_3050_4e07_b3f8_0110308ddcf6.slice - libcontainer container kubepods-besteffort-pode7cdfd74_3050_4e07_b3f8_0110308ddcf6.slice. Nov 24 00:15:30.199556 containerd[1537]: time="2025-11-24T00:15:30.199511754Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5jkmb,Uid:e7cdfd74-3050-4e07-b3f8-0110308ddcf6,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:30.448283 containerd[1537]: time="2025-11-24T00:15:30.448127467Z" level=error msg="Failed to destroy network for sandbox \"8a04ff0649c96001e1eb7f88875f1ce087f6d6fe31b8f41463997d6879cde4b4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:30.451306 systemd[1]: run-netns-cni\x2dfe0125fc\x2d991f\x2d101d\x2dfbad\x2d636c81c5cc8e.mount: Deactivated successfully. Nov 24 00:15:30.522152 containerd[1537]: time="2025-11-24T00:15:30.522058341Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5jkmb,Uid:e7cdfd74-3050-4e07-b3f8-0110308ddcf6,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a04ff0649c96001e1eb7f88875f1ce087f6d6fe31b8f41463997d6879cde4b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:30.522410 kubelet[2683]: E1124 00:15:30.522350 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a04ff0649c96001e1eb7f88875f1ce087f6d6fe31b8f41463997d6879cde4b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:30.522881 kubelet[2683]: E1124 00:15:30.522428 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a04ff0649c96001e1eb7f88875f1ce087f6d6fe31b8f41463997d6879cde4b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:30.522881 kubelet[2683]: E1124 00:15:30.522460 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a04ff0649c96001e1eb7f88875f1ce087f6d6fe31b8f41463997d6879cde4b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:30.522881 kubelet[2683]: E1124 00:15:30.522525 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8a04ff0649c96001e1eb7f88875f1ce087f6d6fe31b8f41463997d6879cde4b4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:36.592057 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount249264120.mount: Deactivated successfully. Nov 24 00:15:41.594342 kubelet[2683]: E1124 00:15:41.594288 2683 kubelet.go:2573] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.404s" Nov 24 00:15:41.597701 containerd[1537]: time="2025-11-24T00:15:41.597658565Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5889ff9675-8zv2t,Uid:024a87f6-fa17-474a-b3f9-39d35fa916e3,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:42.190802 containerd[1537]: time="2025-11-24T00:15:42.190740159Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5jkmb,Uid:e7cdfd74-3050-4e07-b3f8-0110308ddcf6,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:42.967304 containerd[1537]: time="2025-11-24T00:15:42.967202690Z" level=error msg="Failed to destroy network for sandbox \"8162078ab1ebd1784c93ae6d2076d94d9117a1c433658fd50d7148464ea3916c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:42.972145 systemd[1]: run-netns-cni\x2db530ec80\x2d553f\x2d55b2\x2dc1b9\x2dfd05a03c3c47.mount: Deactivated successfully. Nov 24 00:15:43.004569 containerd[1537]: time="2025-11-24T00:15:43.004485956Z" level=error msg="Failed to destroy network for sandbox \"4cc2e0cdfffc35b7b9e7eab546e4a09b8f4d43b5940d6728ca40a2e9af1c9d4b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:43.006940 systemd[1]: run-netns-cni\x2d268569e7\x2dcd80\x2d7706\x2dd0e5\x2dcd7e9e1347de.mount: Deactivated successfully. Nov 24 00:15:43.040050 containerd[1537]: time="2025-11-24T00:15:43.039990463Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:43.143021 containerd[1537]: time="2025-11-24T00:15:43.142912445Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5889ff9675-8zv2t,Uid:024a87f6-fa17-474a-b3f9-39d35fa916e3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8162078ab1ebd1784c93ae6d2076d94d9117a1c433658fd50d7148464ea3916c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:43.143421 kubelet[2683]: E1124 00:15:43.143345 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8162078ab1ebd1784c93ae6d2076d94d9117a1c433658fd50d7148464ea3916c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:43.143876 kubelet[2683]: E1124 00:15:43.143440 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8162078ab1ebd1784c93ae6d2076d94d9117a1c433658fd50d7148464ea3916c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" Nov 24 00:15:43.143876 kubelet[2683]: E1124 00:15:43.143469 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8162078ab1ebd1784c93ae6d2076d94d9117a1c433658fd50d7148464ea3916c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" Nov 24 00:15:43.143876 kubelet[2683]: E1124 00:15:43.143534 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5889ff9675-8zv2t_calico-system(024a87f6-fa17-474a-b3f9-39d35fa916e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5889ff9675-8zv2t_calico-system(024a87f6-fa17-474a-b3f9-39d35fa916e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8162078ab1ebd1784c93ae6d2076d94d9117a1c433658fd50d7148464ea3916c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" podUID="024a87f6-fa17-474a-b3f9-39d35fa916e3" Nov 24 00:15:43.191181 kubelet[2683]: E1124 00:15:43.191061 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:43.191513 containerd[1537]: time="2025-11-24T00:15:43.191470956Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-hr2wl,Uid:22c38642-d71a-4cc4-996b-dcdaded5d62c,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:43.191617 containerd[1537]: time="2025-11-24T00:15:43.191512596Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-n2448,Uid:6cfea110-fbd7-44cd-904d-4c7d268ca542,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:15:43.191750 containerd[1537]: time="2025-11-24T00:15:43.191721731Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-9w4qh,Uid:3b61fcb5-22ca-4d05-817b-ade781b1a990,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:15:43.191819 containerd[1537]: time="2025-11-24T00:15:43.191793500Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-69d76575d7-wx7rs,Uid:bbc561c8-9195-483f-8404-7885126b6cd1,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:43.191876 containerd[1537]: time="2025-11-24T00:15:43.191834569Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-csrvc,Uid:6978f440-5fc4-460d-8ab7-a825799d091f,Namespace:kube-system,Attempt:0,}" Nov 24 00:15:43.196282 containerd[1537]: time="2025-11-24T00:15:43.196178022Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5jkmb,Uid:e7cdfd74-3050-4e07-b3f8-0110308ddcf6,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4cc2e0cdfffc35b7b9e7eab546e4a09b8f4d43b5940d6728ca40a2e9af1c9d4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:43.196504 kubelet[2683]: E1124 00:15:43.196463 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4cc2e0cdfffc35b7b9e7eab546e4a09b8f4d43b5940d6728ca40a2e9af1c9d4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:43.196579 kubelet[2683]: E1124 00:15:43.196523 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4cc2e0cdfffc35b7b9e7eab546e4a09b8f4d43b5940d6728ca40a2e9af1c9d4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:43.196579 kubelet[2683]: E1124 00:15:43.196564 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4cc2e0cdfffc35b7b9e7eab546e4a09b8f4d43b5940d6728ca40a2e9af1c9d4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5jkmb" Nov 24 00:15:43.196684 kubelet[2683]: E1124 00:15:43.196613 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4cc2e0cdfffc35b7b9e7eab546e4a09b8f4d43b5940d6728ca40a2e9af1c9d4b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:15:43.236233 containerd[1537]: time="2025-11-24T00:15:43.236058355Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156883675" Nov 24 00:15:43.886323 containerd[1537]: time="2025-11-24T00:15:43.886185986Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:44.026162 containerd[1537]: time="2025-11-24T00:15:44.026070866Z" level=error msg="Failed to destroy network for sandbox \"6e35bc6e78dc21903ef25f83269af3291b555a314e1a155f8556b9f6dce75295\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.028543 systemd[1]: run-netns-cni\x2dc3b8a37d\x2d44b8\x2de79d\x2dc2b3\x2d5d7e83619844.mount: Deactivated successfully. Nov 24 00:15:44.245814 containerd[1537]: time="2025-11-24T00:15:44.245653458Z" level=error msg="Failed to destroy network for sandbox \"f99123c7d9266ab3284e393f77893cebfad3f128837b3b151aa82cf53d8a3b06\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.404337 containerd[1537]: time="2025-11-24T00:15:44.404273724Z" level=error msg="Failed to destroy network for sandbox \"4e321d2aee49ccc8973ead7a298c16451bdcbd17daa814643d75ddf346de60b9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.434919 systemd[1]: run-netns-cni\x2d3c38e57b\x2d8009\x2db6f0\x2d66cf\x2dffe90786fb43.mount: Deactivated successfully. Nov 24 00:15:44.435039 systemd[1]: run-netns-cni\x2d9945b3de\x2dcc1b\x2d715f\x2da43f\x2d17d7c23e65ae.mount: Deactivated successfully. Nov 24 00:15:44.629838 containerd[1537]: time="2025-11-24T00:15:44.629766581Z" level=error msg="Failed to destroy network for sandbox \"f6237c17600ef18f2c8581f5891f482ffe7b49dadbabf8ffe5888d66ba3616db\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.633017 systemd[1]: run-netns-cni\x2d6b601b11\x2df2a5\x2d0697\x2d629b\x2d9495fd7ee369.mount: Deactivated successfully. Nov 24 00:15:44.725464 containerd[1537]: time="2025-11-24T00:15:44.725377101Z" level=error msg="Failed to destroy network for sandbox \"2c995872b8d481877b0531bf7f33d307b5610195e7be444c4e7a111497ee9732\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.728341 systemd[1]: run-netns-cni\x2d03dc5d5b\x2dbe8c\x2d564c\x2d6d48\x2d4605ef9d4b8f.mount: Deactivated successfully. Nov 24 00:15:44.784137 containerd[1537]: time="2025-11-24T00:15:44.784054943Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-hr2wl,Uid:22c38642-d71a-4cc4-996b-dcdaded5d62c,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"6e35bc6e78dc21903ef25f83269af3291b555a314e1a155f8556b9f6dce75295\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.784474 kubelet[2683]: E1124 00:15:44.784420 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6e35bc6e78dc21903ef25f83269af3291b555a314e1a155f8556b9f6dce75295\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.784853 kubelet[2683]: E1124 00:15:44.784503 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6e35bc6e78dc21903ef25f83269af3291b555a314e1a155f8556b9f6dce75295\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-hr2wl" Nov 24 00:15:44.784853 kubelet[2683]: E1124 00:15:44.784525 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6e35bc6e78dc21903ef25f83269af3291b555a314e1a155f8556b9f6dce75295\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-hr2wl" Nov 24 00:15:44.784853 kubelet[2683]: E1124 00:15:44.784586 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-hr2wl_calico-system(22c38642-d71a-4cc4-996b-dcdaded5d62c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-hr2wl_calico-system(22c38642-d71a-4cc4-996b-dcdaded5d62c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6e35bc6e78dc21903ef25f83269af3291b555a314e1a155f8556b9f6dce75295\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-hr2wl" podUID="22c38642-d71a-4cc4-996b-dcdaded5d62c" Nov 24 00:15:44.869810 systemd[1]: Started sshd@7-10.0.0.25:22-10.0.0.1:40340.service - OpenSSH per-connection server daemon (10.0.0.1:40340). Nov 24 00:15:44.958285 containerd[1537]: time="2025-11-24T00:15:44.958195188Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-9w4qh,Uid:3b61fcb5-22ca-4d05-817b-ade781b1a990,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4e321d2aee49ccc8973ead7a298c16451bdcbd17daa814643d75ddf346de60b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.958653 kubelet[2683]: E1124 00:15:44.958571 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4e321d2aee49ccc8973ead7a298c16451bdcbd17daa814643d75ddf346de60b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:44.958717 kubelet[2683]: E1124 00:15:44.958671 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4e321d2aee49ccc8973ead7a298c16451bdcbd17daa814643d75ddf346de60b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" Nov 24 00:15:44.958717 kubelet[2683]: E1124 00:15:44.958707 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4e321d2aee49ccc8973ead7a298c16451bdcbd17daa814643d75ddf346de60b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" Nov 24 00:15:44.958788 kubelet[2683]: E1124 00:15:44.958756 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c779694df-9w4qh_calico-apiserver(3b61fcb5-22ca-4d05-817b-ade781b1a990)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c779694df-9w4qh_calico-apiserver(3b61fcb5-22ca-4d05-817b-ade781b1a990)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4e321d2aee49ccc8973ead7a298c16451bdcbd17daa814643d75ddf346de60b9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" podUID="3b61fcb5-22ca-4d05-817b-ade781b1a990" Nov 24 00:15:45.006558 sshd[4026]: Accepted publickey for core from 10.0.0.1 port 40340 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:15:45.008571 sshd-session[4026]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:15:45.019978 systemd-logind[1519]: New session 8 of user core. Nov 24 00:15:45.028799 systemd[1]: Started session-8.scope - Session 8 of User core. Nov 24 00:15:45.041649 containerd[1537]: time="2025-11-24T00:15:45.041548713Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-69d76575d7-wx7rs,Uid:bbc561c8-9195-483f-8404-7885126b6cd1,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f6237c17600ef18f2c8581f5891f482ffe7b49dadbabf8ffe5888d66ba3616db\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.042081 kubelet[2683]: E1124 00:15:45.041835 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f6237c17600ef18f2c8581f5891f482ffe7b49dadbabf8ffe5888d66ba3616db\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.042081 kubelet[2683]: E1124 00:15:45.041901 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f6237c17600ef18f2c8581f5891f482ffe7b49dadbabf8ffe5888d66ba3616db\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-69d76575d7-wx7rs" Nov 24 00:15:45.042081 kubelet[2683]: E1124 00:15:45.041923 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f6237c17600ef18f2c8581f5891f482ffe7b49dadbabf8ffe5888d66ba3616db\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-69d76575d7-wx7rs" Nov 24 00:15:45.042184 kubelet[2683]: E1124 00:15:45.041971 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-69d76575d7-wx7rs_calico-system(bbc561c8-9195-483f-8404-7885126b6cd1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-69d76575d7-wx7rs_calico-system(bbc561c8-9195-483f-8404-7885126b6cd1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f6237c17600ef18f2c8581f5891f482ffe7b49dadbabf8ffe5888d66ba3616db\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-69d76575d7-wx7rs" podUID="bbc561c8-9195-483f-8404-7885126b6cd1" Nov 24 00:15:45.073213 containerd[1537]: time="2025-11-24T00:15:45.073129727Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 24 00:15:45.073759 containerd[1537]: time="2025-11-24T00:15:45.073714787Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 15.746501328s" Nov 24 00:15:45.073806 containerd[1537]: time="2025-11-24T00:15:45.073763320Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Nov 24 00:15:45.085377 containerd[1537]: time="2025-11-24T00:15:45.085332075Z" level=info msg="CreateContainer within sandbox \"690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Nov 24 00:15:45.155981 containerd[1537]: time="2025-11-24T00:15:45.155912631Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-csrvc,Uid:6978f440-5fc4-460d-8ab7-a825799d091f,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c995872b8d481877b0531bf7f33d307b5610195e7be444c4e7a111497ee9732\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.156227 kubelet[2683]: E1124 00:15:45.156180 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c995872b8d481877b0531bf7f33d307b5610195e7be444c4e7a111497ee9732\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.156311 kubelet[2683]: E1124 00:15:45.156257 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c995872b8d481877b0531bf7f33d307b5610195e7be444c4e7a111497ee9732\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-csrvc" Nov 24 00:15:45.156311 kubelet[2683]: E1124 00:15:45.156291 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c995872b8d481877b0531bf7f33d307b5610195e7be444c4e7a111497ee9732\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-csrvc" Nov 24 00:15:45.156410 kubelet[2683]: E1124 00:15:45.156347 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-csrvc_kube-system(6978f440-5fc4-460d-8ab7-a825799d091f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-csrvc_kube-system(6978f440-5fc4-460d-8ab7-a825799d091f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2c995872b8d481877b0531bf7f33d307b5610195e7be444c4e7a111497ee9732\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-csrvc" podUID="6978f440-5fc4-460d-8ab7-a825799d091f" Nov 24 00:15:45.192077 kubelet[2683]: E1124 00:15:45.191952 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:45.194307 containerd[1537]: time="2025-11-24T00:15:45.193908805Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fskrk,Uid:d27f4aa2-b9ea-448e-a394-15df266e8761,Namespace:kube-system,Attempt:0,}" Nov 24 00:15:45.196518 containerd[1537]: time="2025-11-24T00:15:45.196334400Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-n2448,Uid:6cfea110-fbd7-44cd-904d-4c7d268ca542,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f99123c7d9266ab3284e393f77893cebfad3f128837b3b151aa82cf53d8a3b06\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.197031 kubelet[2683]: E1124 00:15:45.196967 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f99123c7d9266ab3284e393f77893cebfad3f128837b3b151aa82cf53d8a3b06\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.197109 kubelet[2683]: E1124 00:15:45.197038 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f99123c7d9266ab3284e393f77893cebfad3f128837b3b151aa82cf53d8a3b06\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c779694df-n2448" Nov 24 00:15:45.197109 kubelet[2683]: E1124 00:15:45.197065 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f99123c7d9266ab3284e393f77893cebfad3f128837b3b151aa82cf53d8a3b06\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c779694df-n2448" Nov 24 00:15:45.197184 kubelet[2683]: E1124 00:15:45.197113 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c779694df-n2448_calico-apiserver(6cfea110-fbd7-44cd-904d-4c7d268ca542)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c779694df-n2448_calico-apiserver(6cfea110-fbd7-44cd-904d-4c7d268ca542)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f99123c7d9266ab3284e393f77893cebfad3f128837b3b151aa82cf53d8a3b06\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c779694df-n2448" podUID="6cfea110-fbd7-44cd-904d-4c7d268ca542" Nov 24 00:15:45.341578 sshd[4029]: Connection closed by 10.0.0.1 port 40340 Nov 24 00:15:45.343822 sshd-session[4026]: pam_unix(sshd:session): session closed for user core Nov 24 00:15:45.349377 systemd[1]: sshd@7-10.0.0.25:22-10.0.0.1:40340.service: Deactivated successfully. Nov 24 00:15:45.351694 systemd[1]: session-8.scope: Deactivated successfully. Nov 24 00:15:45.352661 systemd-logind[1519]: Session 8 logged out. Waiting for processes to exit. Nov 24 00:15:45.354409 systemd-logind[1519]: Removed session 8. Nov 24 00:15:45.384928 containerd[1537]: time="2025-11-24T00:15:45.384845754Z" level=info msg="Container f5088f742cb883dc54ed03befb080dc63c81a1cbaebf1eefb89dc1f61b7beb65: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:15:45.389337 containerd[1537]: time="2025-11-24T00:15:45.389272645Z" level=error msg="Failed to destroy network for sandbox \"062623377bcca441fbcddae0607a4d8564fb750cebb18dd20388e7d8e9e8e480\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.484337 containerd[1537]: time="2025-11-24T00:15:45.484244901Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fskrk,Uid:d27f4aa2-b9ea-448e-a394-15df266e8761,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"062623377bcca441fbcddae0607a4d8564fb750cebb18dd20388e7d8e9e8e480\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.484697 kubelet[2683]: E1124 00:15:45.484610 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"062623377bcca441fbcddae0607a4d8564fb750cebb18dd20388e7d8e9e8e480\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 24 00:15:45.484793 kubelet[2683]: E1124 00:15:45.484724 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"062623377bcca441fbcddae0607a4d8564fb750cebb18dd20388e7d8e9e8e480\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-fskrk" Nov 24 00:15:45.484793 kubelet[2683]: E1124 00:15:45.484747 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"062623377bcca441fbcddae0607a4d8564fb750cebb18dd20388e7d8e9e8e480\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-fskrk" Nov 24 00:15:45.484870 kubelet[2683]: E1124 00:15:45.484796 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-fskrk_kube-system(d27f4aa2-b9ea-448e-a394-15df266e8761)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-fskrk_kube-system(d27f4aa2-b9ea-448e-a394-15df266e8761)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"062623377bcca441fbcddae0607a4d8564fb750cebb18dd20388e7d8e9e8e480\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-fskrk" podUID="d27f4aa2-b9ea-448e-a394-15df266e8761" Nov 24 00:15:45.964177 containerd[1537]: time="2025-11-24T00:15:45.964098231Z" level=info msg="CreateContainer within sandbox \"690fbbf79e9e37942728ebe7c60e1b9110e1568c434fd22103e1b57295226a26\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"f5088f742cb883dc54ed03befb080dc63c81a1cbaebf1eefb89dc1f61b7beb65\"" Nov 24 00:15:45.964976 containerd[1537]: time="2025-11-24T00:15:45.964922884Z" level=info msg="StartContainer for \"f5088f742cb883dc54ed03befb080dc63c81a1cbaebf1eefb89dc1f61b7beb65\"" Nov 24 00:15:45.966865 containerd[1537]: time="2025-11-24T00:15:45.966773979Z" level=info msg="connecting to shim f5088f742cb883dc54ed03befb080dc63c81a1cbaebf1eefb89dc1f61b7beb65" address="unix:///run/containerd/s/7f7049f7fa2eddffbd1af32ef80cfeea7699eafe3d373bcadc21d6e309b3d52d" protocol=ttrpc version=3 Nov 24 00:15:46.000928 systemd[1]: Started cri-containerd-f5088f742cb883dc54ed03befb080dc63c81a1cbaebf1eefb89dc1f61b7beb65.scope - libcontainer container f5088f742cb883dc54ed03befb080dc63c81a1cbaebf1eefb89dc1f61b7beb65. Nov 24 00:15:46.184508 containerd[1537]: time="2025-11-24T00:15:46.184381698Z" level=info msg="StartContainer for \"f5088f742cb883dc54ed03befb080dc63c81a1cbaebf1eefb89dc1f61b7beb65\" returns successfully" Nov 24 00:15:46.223280 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Nov 24 00:15:46.223436 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Nov 24 00:15:46.619167 kubelet[2683]: E1124 00:15:46.619116 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:47.020887 kubelet[2683]: I1124 00:15:47.020695 2683 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbc561c8-9195-483f-8404-7885126b6cd1-whisker-ca-bundle\") pod \"bbc561c8-9195-483f-8404-7885126b6cd1\" (UID: \"bbc561c8-9195-483f-8404-7885126b6cd1\") " Nov 24 00:15:47.020887 kubelet[2683]: I1124 00:15:47.020786 2683 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vllsh\" (UniqueName: \"kubernetes.io/projected/bbc561c8-9195-483f-8404-7885126b6cd1-kube-api-access-vllsh\") pod \"bbc561c8-9195-483f-8404-7885126b6cd1\" (UID: \"bbc561c8-9195-483f-8404-7885126b6cd1\") " Nov 24 00:15:47.020887 kubelet[2683]: I1124 00:15:47.020820 2683 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/bbc561c8-9195-483f-8404-7885126b6cd1-whisker-backend-key-pair\") pod \"bbc561c8-9195-483f-8404-7885126b6cd1\" (UID: \"bbc561c8-9195-483f-8404-7885126b6cd1\") " Nov 24 00:15:47.021400 kubelet[2683]: I1124 00:15:47.021336 2683 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbc561c8-9195-483f-8404-7885126b6cd1-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "bbc561c8-9195-483f-8404-7885126b6cd1" (UID: "bbc561c8-9195-483f-8404-7885126b6cd1"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Nov 24 00:15:47.026461 kubelet[2683]: I1124 00:15:47.026408 2683 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc561c8-9195-483f-8404-7885126b6cd1-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "bbc561c8-9195-483f-8404-7885126b6cd1" (UID: "bbc561c8-9195-483f-8404-7885126b6cd1"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Nov 24 00:15:47.026615 kubelet[2683]: I1124 00:15:47.026588 2683 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbc561c8-9195-483f-8404-7885126b6cd1-kube-api-access-vllsh" (OuterVolumeSpecName: "kube-api-access-vllsh") pod "bbc561c8-9195-483f-8404-7885126b6cd1" (UID: "bbc561c8-9195-483f-8404-7885126b6cd1"). InnerVolumeSpecName "kube-api-access-vllsh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Nov 24 00:15:47.027464 systemd[1]: var-lib-kubelet-pods-bbc561c8\x2d9195\x2d483f\x2d8404\x2d7885126b6cd1-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dvllsh.mount: Deactivated successfully. Nov 24 00:15:47.027609 systemd[1]: var-lib-kubelet-pods-bbc561c8\x2d9195\x2d483f\x2d8404\x2d7885126b6cd1-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Nov 24 00:15:47.100708 kubelet[2683]: I1124 00:15:47.100232 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-nh7wm" podStartSLOduration=3.313334738 podStartE2EDuration="33.100202991s" podCreationTimestamp="2025-11-24 00:15:14 +0000 UTC" firstStartedPulling="2025-11-24 00:15:15.287679883 +0000 UTC m=+30.195325235" lastFinishedPulling="2025-11-24 00:15:45.074548136 +0000 UTC m=+59.982193488" observedRunningTime="2025-11-24 00:15:47.098466623 +0000 UTC m=+62.006111985" watchObservedRunningTime="2025-11-24 00:15:47.100202991 +0000 UTC m=+62.007848343" Nov 24 00:15:47.122002 kubelet[2683]: I1124 00:15:47.121931 2683 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vllsh\" (UniqueName: \"kubernetes.io/projected/bbc561c8-9195-483f-8404-7885126b6cd1-kube-api-access-vllsh\") on node \"localhost\" DevicePath \"\"" Nov 24 00:15:47.122002 kubelet[2683]: I1124 00:15:47.121971 2683 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/bbc561c8-9195-483f-8404-7885126b6cd1-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Nov 24 00:15:47.122002 kubelet[2683]: I1124 00:15:47.121983 2683 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbc561c8-9195-483f-8404-7885126b6cd1-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Nov 24 00:15:47.198707 systemd[1]: Removed slice kubepods-besteffort-podbbc561c8_9195_483f_8404_7885126b6cd1.slice - libcontainer container kubepods-besteffort-podbbc561c8_9195_483f_8404_7885126b6cd1.slice. Nov 24 00:15:47.621292 kubelet[2683]: E1124 00:15:47.621230 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:48.697971 systemd[1]: Created slice kubepods-besteffort-pod4aa1d061_de13_4af8_bf9e_bd05bcbcd8d2.slice - libcontainer container kubepods-besteffort-pod4aa1d061_de13_4af8_bf9e_bd05bcbcd8d2.slice. Nov 24 00:15:48.834381 kubelet[2683]: I1124 00:15:48.834281 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wwgj\" (UniqueName: \"kubernetes.io/projected/4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2-kube-api-access-4wwgj\") pod \"whisker-7657d947d7-4pbrs\" (UID: \"4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2\") " pod="calico-system/whisker-7657d947d7-4pbrs" Nov 24 00:15:48.834381 kubelet[2683]: I1124 00:15:48.834360 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2-whisker-ca-bundle\") pod \"whisker-7657d947d7-4pbrs\" (UID: \"4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2\") " pod="calico-system/whisker-7657d947d7-4pbrs" Nov 24 00:15:48.835124 kubelet[2683]: I1124 00:15:48.834397 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2-whisker-backend-key-pair\") pod \"whisker-7657d947d7-4pbrs\" (UID: \"4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2\") " pod="calico-system/whisker-7657d947d7-4pbrs" Nov 24 00:15:49.004199 containerd[1537]: time="2025-11-24T00:15:49.003977458Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7657d947d7-4pbrs,Uid:4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:49.193314 kubelet[2683]: I1124 00:15:49.193246 2683 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbc561c8-9195-483f-8404-7885126b6cd1" path="/var/lib/kubelet/pods/bbc561c8-9195-483f-8404-7885126b6cd1/volumes" Nov 24 00:15:49.351018 systemd-networkd[1448]: cali5c284b77720: Link UP Nov 24 00:15:49.353917 systemd-networkd[1448]: cali5c284b77720: Gained carrier Nov 24 00:15:49.379504 containerd[1537]: 2025-11-24 00:15:49.137 [INFO][4210] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Nov 24 00:15:49.379504 containerd[1537]: 2025-11-24 00:15:49.171 [INFO][4210] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--7657d947d7--4pbrs-eth0 whisker-7657d947d7- calico-system 4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2 983 0 2025-11-24 00:15:48 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:7657d947d7 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-7657d947d7-4pbrs eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali5c284b77720 [] [] }} ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Namespace="calico-system" Pod="whisker-7657d947d7-4pbrs" WorkloadEndpoint="localhost-k8s-whisker--7657d947d7--4pbrs-" Nov 24 00:15:49.379504 containerd[1537]: 2025-11-24 00:15:49.171 [INFO][4210] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Namespace="calico-system" Pod="whisker-7657d947d7-4pbrs" WorkloadEndpoint="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" Nov 24 00:15:49.379504 containerd[1537]: 2025-11-24 00:15:49.280 [INFO][4218] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" HandleID="k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Workload="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.283 [INFO][4218] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" HandleID="k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Workload="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000124320), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-7657d947d7-4pbrs", "timestamp":"2025-11-24 00:15:49.280500411 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.283 [INFO][4218] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.283 [INFO][4218] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.284 [INFO][4218] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.295 [INFO][4218] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" host="localhost" Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.304 [INFO][4218] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.309 [INFO][4218] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.311 [INFO][4218] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.313 [INFO][4218] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:49.379796 containerd[1537]: 2025-11-24 00:15:49.313 [INFO][4218] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" host="localhost" Nov 24 00:15:49.380212 containerd[1537]: 2025-11-24 00:15:49.315 [INFO][4218] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de Nov 24 00:15:49.380212 containerd[1537]: 2025-11-24 00:15:49.319 [INFO][4218] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" host="localhost" Nov 24 00:15:49.380212 containerd[1537]: 2025-11-24 00:15:49.327 [INFO][4218] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" host="localhost" Nov 24 00:15:49.380212 containerd[1537]: 2025-11-24 00:15:49.327 [INFO][4218] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" host="localhost" Nov 24 00:15:49.380212 containerd[1537]: 2025-11-24 00:15:49.327 [INFO][4218] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:15:49.380212 containerd[1537]: 2025-11-24 00:15:49.327 [INFO][4218] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" HandleID="k8s-pod-network.24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Workload="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" Nov 24 00:15:49.380354 containerd[1537]: 2025-11-24 00:15:49.332 [INFO][4210] cni-plugin/k8s.go 418: Populated endpoint ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Namespace="calico-system" Pod="whisker-7657d947d7-4pbrs" WorkloadEndpoint="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--7657d947d7--4pbrs-eth0", GenerateName:"whisker-7657d947d7-", Namespace:"calico-system", SelfLink:"", UID:"4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2", ResourceVersion:"983", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7657d947d7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-7657d947d7-4pbrs", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali5c284b77720", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:49.380354 containerd[1537]: 2025-11-24 00:15:49.333 [INFO][4210] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Namespace="calico-system" Pod="whisker-7657d947d7-4pbrs" WorkloadEndpoint="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" Nov 24 00:15:49.380471 containerd[1537]: 2025-11-24 00:15:49.333 [INFO][4210] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5c284b77720 ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Namespace="calico-system" Pod="whisker-7657d947d7-4pbrs" WorkloadEndpoint="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" Nov 24 00:15:49.380471 containerd[1537]: 2025-11-24 00:15:49.359 [INFO][4210] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Namespace="calico-system" Pod="whisker-7657d947d7-4pbrs" WorkloadEndpoint="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" Nov 24 00:15:49.380523 containerd[1537]: 2025-11-24 00:15:49.361 [INFO][4210] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Namespace="calico-system" Pod="whisker-7657d947d7-4pbrs" WorkloadEndpoint="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--7657d947d7--4pbrs-eth0", GenerateName:"whisker-7657d947d7-", Namespace:"calico-system", SelfLink:"", UID:"4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2", ResourceVersion:"983", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7657d947d7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de", Pod:"whisker-7657d947d7-4pbrs", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali5c284b77720", MAC:"d2:a6:27:42:7c:72", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:49.380583 containerd[1537]: 2025-11-24 00:15:49.373 [INFO][4210] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" Namespace="calico-system" Pod="whisker-7657d947d7-4pbrs" WorkloadEndpoint="localhost-k8s-whisker--7657d947d7--4pbrs-eth0" Nov 24 00:15:49.522951 containerd[1537]: time="2025-11-24T00:15:49.522878759Z" level=info msg="connecting to shim 24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de" address="unix:///run/containerd/s/a5cc163a03e2c10605fd9d7ef3591ad6991833a37d067d09cf69145538a82104" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:49.574120 systemd[1]: Started cri-containerd-24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de.scope - libcontainer container 24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de. Nov 24 00:15:49.597430 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 24 00:15:49.845465 containerd[1537]: time="2025-11-24T00:15:49.845403538Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7657d947d7-4pbrs,Uid:4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2,Namespace:calico-system,Attempt:0,} returns sandbox id \"24b87e82888cd24760a86e6c064d5a0e02bbb5fd435e283d7fd7f0d6a64ab1de\"" Nov 24 00:15:49.847220 containerd[1537]: time="2025-11-24T00:15:49.847040019Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 24 00:15:49.961315 systemd-networkd[1448]: vxlan.calico: Link UP Nov 24 00:15:49.961327 systemd-networkd[1448]: vxlan.calico: Gained carrier Nov 24 00:15:50.268771 containerd[1537]: time="2025-11-24T00:15:50.268556633Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:15:50.272330 containerd[1537]: time="2025-11-24T00:15:50.271165452Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 24 00:15:50.282326 containerd[1537]: time="2025-11-24T00:15:50.282207429Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 24 00:15:50.282718 kubelet[2683]: E1124 00:15:50.282617 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:15:50.285197 kubelet[2683]: E1124 00:15:50.283136 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:15:50.297907 kubelet[2683]: E1124 00:15:50.297644 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:ed3f3941752e478da85a0af9ae40cdb3,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4wwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-7657d947d7-4pbrs_calico-system(4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 24 00:15:50.301650 containerd[1537]: time="2025-11-24T00:15:50.301573471Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 24 00:15:50.358558 systemd[1]: Started sshd@8-10.0.0.25:22-10.0.0.1:39364.service - OpenSSH per-connection server daemon (10.0.0.1:39364). Nov 24 00:15:50.426465 sshd[4482]: Accepted publickey for core from 10.0.0.1 port 39364 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:15:50.428516 sshd-session[4482]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:15:50.435205 systemd-logind[1519]: New session 9 of user core. Nov 24 00:15:50.446003 systemd[1]: Started session-9.scope - Session 9 of User core. Nov 24 00:15:50.556962 systemd-networkd[1448]: cali5c284b77720: Gained IPv6LL Nov 24 00:15:50.600304 sshd[4486]: Connection closed by 10.0.0.1 port 39364 Nov 24 00:15:50.600818 sshd-session[4482]: pam_unix(sshd:session): session closed for user core Nov 24 00:15:50.606198 systemd[1]: sshd@8-10.0.0.25:22-10.0.0.1:39364.service: Deactivated successfully. Nov 24 00:15:50.608886 systemd[1]: session-9.scope: Deactivated successfully. Nov 24 00:15:50.609894 systemd-logind[1519]: Session 9 logged out. Waiting for processes to exit. Nov 24 00:15:50.611713 systemd-logind[1519]: Removed session 9. Nov 24 00:15:50.643329 containerd[1537]: time="2025-11-24T00:15:50.643264371Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:15:50.729915 containerd[1537]: time="2025-11-24T00:15:50.729789348Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 24 00:15:50.729915 containerd[1537]: time="2025-11-24T00:15:50.729900482Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 24 00:15:50.730199 kubelet[2683]: E1124 00:15:50.730142 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:15:50.730272 kubelet[2683]: E1124 00:15:50.730210 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:15:50.730435 kubelet[2683]: E1124 00:15:50.730362 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-7657d947d7-4pbrs_calico-system(4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 24 00:15:50.731681 kubelet[2683]: E1124 00:15:50.731610 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-7657d947d7-4pbrs" podUID="4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2" Nov 24 00:15:51.196829 systemd-networkd[1448]: vxlan.calico: Gained IPv6LL Nov 24 00:15:51.632917 kubelet[2683]: E1124 00:15:51.632813 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-7657d947d7-4pbrs" podUID="4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2" Nov 24 00:15:55.191583 containerd[1537]: time="2025-11-24T00:15:55.191505340Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5889ff9675-8zv2t,Uid:024a87f6-fa17-474a-b3f9-39d35fa916e3,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:55.312294 systemd-networkd[1448]: calid18c658d033: Link UP Nov 24 00:15:55.313381 systemd-networkd[1448]: calid18c658d033: Gained carrier Nov 24 00:15:55.521316 containerd[1537]: 2025-11-24 00:15:55.237 [INFO][4507] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0 calico-kube-controllers-5889ff9675- calico-system 024a87f6-fa17-474a-b3f9-39d35fa916e3 854 0 2025-11-24 00:15:14 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5889ff9675 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-5889ff9675-8zv2t eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calid18c658d033 [] [] }} ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Namespace="calico-system" Pod="calico-kube-controllers-5889ff9675-8zv2t" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-" Nov 24 00:15:55.521316 containerd[1537]: 2025-11-24 00:15:55.237 [INFO][4507] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Namespace="calico-system" Pod="calico-kube-controllers-5889ff9675-8zv2t" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" Nov 24 00:15:55.521316 containerd[1537]: 2025-11-24 00:15:55.266 [INFO][4522] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" HandleID="k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Workload="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.266 [INFO][4522] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" HandleID="k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Workload="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001393f0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-5889ff9675-8zv2t", "timestamp":"2025-11-24 00:15:55.266690814 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.266 [INFO][4522] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.266 [INFO][4522] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.266 [INFO][4522] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.276 [INFO][4522] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" host="localhost" Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.281 [INFO][4522] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.285 [INFO][4522] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.287 [INFO][4522] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.290 [INFO][4522] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:55.521739 containerd[1537]: 2025-11-24 00:15:55.290 [INFO][4522] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" host="localhost" Nov 24 00:15:55.522392 containerd[1537]: 2025-11-24 00:15:55.292 [INFO][4522] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549 Nov 24 00:15:55.522392 containerd[1537]: 2025-11-24 00:15:55.296 [INFO][4522] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" host="localhost" Nov 24 00:15:55.522392 containerd[1537]: 2025-11-24 00:15:55.305 [INFO][4522] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" host="localhost" Nov 24 00:15:55.522392 containerd[1537]: 2025-11-24 00:15:55.305 [INFO][4522] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" host="localhost" Nov 24 00:15:55.522392 containerd[1537]: 2025-11-24 00:15:55.305 [INFO][4522] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:15:55.522392 containerd[1537]: 2025-11-24 00:15:55.305 [INFO][4522] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" HandleID="k8s-pod-network.7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Workload="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" Nov 24 00:15:55.522594 containerd[1537]: 2025-11-24 00:15:55.309 [INFO][4507] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Namespace="calico-system" Pod="calico-kube-controllers-5889ff9675-8zv2t" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0", GenerateName:"calico-kube-controllers-5889ff9675-", Namespace:"calico-system", SelfLink:"", UID:"024a87f6-fa17-474a-b3f9-39d35fa916e3", ResourceVersion:"854", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5889ff9675", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-5889ff9675-8zv2t", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calid18c658d033", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:55.522707 containerd[1537]: 2025-11-24 00:15:55.309 [INFO][4507] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Namespace="calico-system" Pod="calico-kube-controllers-5889ff9675-8zv2t" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" Nov 24 00:15:55.522707 containerd[1537]: 2025-11-24 00:15:55.309 [INFO][4507] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid18c658d033 ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Namespace="calico-system" Pod="calico-kube-controllers-5889ff9675-8zv2t" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" Nov 24 00:15:55.522707 containerd[1537]: 2025-11-24 00:15:55.312 [INFO][4507] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Namespace="calico-system" Pod="calico-kube-controllers-5889ff9675-8zv2t" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" Nov 24 00:15:55.522812 containerd[1537]: 2025-11-24 00:15:55.313 [INFO][4507] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Namespace="calico-system" Pod="calico-kube-controllers-5889ff9675-8zv2t" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0", GenerateName:"calico-kube-controllers-5889ff9675-", Namespace:"calico-system", SelfLink:"", UID:"024a87f6-fa17-474a-b3f9-39d35fa916e3", ResourceVersion:"854", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5889ff9675", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549", Pod:"calico-kube-controllers-5889ff9675-8zv2t", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calid18c658d033", MAC:"76:78:8e:27:f3:c8", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:55.522899 containerd[1537]: 2025-11-24 00:15:55.517 [INFO][4507] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" Namespace="calico-system" Pod="calico-kube-controllers-5889ff9675-8zv2t" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5889ff9675--8zv2t-eth0" Nov 24 00:15:55.616269 systemd[1]: Started sshd@9-10.0.0.25:22-10.0.0.1:39380.service - OpenSSH per-connection server daemon (10.0.0.1:39380). Nov 24 00:15:55.682990 sshd[4550]: Accepted publickey for core from 10.0.0.1 port 39380 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:15:55.688742 sshd-session[4550]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:15:55.693051 systemd-logind[1519]: New session 10 of user core. Nov 24 00:15:55.703772 systemd[1]: Started session-10.scope - Session 10 of User core. Nov 24 00:15:55.826644 sshd[4553]: Connection closed by 10.0.0.1 port 39380 Nov 24 00:15:55.827016 sshd-session[4550]: pam_unix(sshd:session): session closed for user core Nov 24 00:15:55.831828 systemd[1]: sshd@9-10.0.0.25:22-10.0.0.1:39380.service: Deactivated successfully. Nov 24 00:15:55.833909 systemd[1]: session-10.scope: Deactivated successfully. Nov 24 00:15:55.834888 systemd-logind[1519]: Session 10 logged out. Waiting for processes to exit. Nov 24 00:15:55.836372 systemd-logind[1519]: Removed session 10. Nov 24 00:15:56.341395 containerd[1537]: time="2025-11-24T00:15:56.341334237Z" level=info msg="connecting to shim 7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549" address="unix:///run/containerd/s/64ac78a86d2b95e8bd9695020496beb6b065f12c6e9b7f6d70bda65af81113f7" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:56.369994 systemd[1]: Started cri-containerd-7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549.scope - libcontainer container 7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549. Nov 24 00:15:56.386708 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 24 00:15:56.464005 containerd[1537]: time="2025-11-24T00:15:56.463907181Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5889ff9675-8zv2t,Uid:024a87f6-fa17-474a-b3f9-39d35fa916e3,Namespace:calico-system,Attempt:0,} returns sandbox id \"7f368be2e633f9505e573aeb889fd0f90352de233eee2c48790cca069cf34549\"" Nov 24 00:15:56.465609 containerd[1537]: time="2025-11-24T00:15:56.465577132Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 24 00:15:56.636806 systemd-networkd[1448]: calid18c658d033: Gained IPv6LL Nov 24 00:15:56.852424 containerd[1537]: time="2025-11-24T00:15:56.852358874Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:15:56.853683 containerd[1537]: time="2025-11-24T00:15:56.853606536Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 24 00:15:56.853683 containerd[1537]: time="2025-11-24T00:15:56.853662894Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 24 00:15:56.854096 kubelet[2683]: E1124 00:15:56.853919 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:15:56.854569 kubelet[2683]: E1124 00:15:56.854106 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:15:56.854569 kubelet[2683]: E1124 00:15:56.854313 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xn2qq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5889ff9675-8zv2t_calico-system(024a87f6-fa17-474a-b3f9-39d35fa916e3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 24 00:15:56.855564 kubelet[2683]: E1124 00:15:56.855497 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" podUID="024a87f6-fa17-474a-b3f9-39d35fa916e3" Nov 24 00:15:57.190980 kubelet[2683]: E1124 00:15:57.190940 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:57.191347 containerd[1537]: time="2025-11-24T00:15:57.191306349Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-csrvc,Uid:6978f440-5fc4-460d-8ab7-a825799d091f,Namespace:kube-system,Attempt:0,}" Nov 24 00:15:57.191894 containerd[1537]: time="2025-11-24T00:15:57.191843057Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-hr2wl,Uid:22c38642-d71a-4cc4-996b-dcdaded5d62c,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:57.192166 containerd[1537]: time="2025-11-24T00:15:57.191860811Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-9w4qh,Uid:3b61fcb5-22ca-4d05-817b-ade781b1a990,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:15:57.604598 systemd-networkd[1448]: califf78334c3e7: Link UP Nov 24 00:15:57.606835 systemd-networkd[1448]: califf78334c3e7: Gained carrier Nov 24 00:15:57.651250 kubelet[2683]: E1124 00:15:57.649679 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" podUID="024a87f6-fa17-474a-b3f9-39d35fa916e3" Nov 24 00:15:57.657453 containerd[1537]: 2025-11-24 00:15:57.441 [INFO][4614] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--csrvc-eth0 coredns-668d6bf9bc- kube-system 6978f440-5fc4-460d-8ab7-a825799d091f 849 0 2025-11-24 00:14:51 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-csrvc eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] califf78334c3e7 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Namespace="kube-system" Pod="coredns-668d6bf9bc-csrvc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--csrvc-" Nov 24 00:15:57.657453 containerd[1537]: 2025-11-24 00:15:57.441 [INFO][4614] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Namespace="kube-system" Pod="coredns-668d6bf9bc-csrvc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" Nov 24 00:15:57.657453 containerd[1537]: 2025-11-24 00:15:57.471 [INFO][4629] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" HandleID="k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Workload="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.471 [INFO][4629] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" HandleID="k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Workload="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00043a050), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-csrvc", "timestamp":"2025-11-24 00:15:57.471535011 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.471 [INFO][4629] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.471 [INFO][4629] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.471 [INFO][4629] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.480 [INFO][4629] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" host="localhost" Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.485 [INFO][4629] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.488 [INFO][4629] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.490 [INFO][4629] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.492 [INFO][4629] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:57.658233 containerd[1537]: 2025-11-24 00:15:57.492 [INFO][4629] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" host="localhost" Nov 24 00:15:57.658544 containerd[1537]: 2025-11-24 00:15:57.493 [INFO][4629] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2 Nov 24 00:15:57.658544 containerd[1537]: 2025-11-24 00:15:57.510 [INFO][4629] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" host="localhost" Nov 24 00:15:57.658544 containerd[1537]: 2025-11-24 00:15:57.597 [INFO][4629] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" host="localhost" Nov 24 00:15:57.658544 containerd[1537]: 2025-11-24 00:15:57.598 [INFO][4629] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" host="localhost" Nov 24 00:15:57.658544 containerd[1537]: 2025-11-24 00:15:57.598 [INFO][4629] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:15:57.658544 containerd[1537]: 2025-11-24 00:15:57.598 [INFO][4629] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" HandleID="k8s-pod-network.34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Workload="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" Nov 24 00:15:57.658751 containerd[1537]: 2025-11-24 00:15:57.601 [INFO][4614] cni-plugin/k8s.go 418: Populated endpoint ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Namespace="kube-system" Pod="coredns-668d6bf9bc-csrvc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--csrvc-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"6978f440-5fc4-460d-8ab7-a825799d091f", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 14, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-csrvc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"califf78334c3e7", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:57.658836 containerd[1537]: 2025-11-24 00:15:57.601 [INFO][4614] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Namespace="kube-system" Pod="coredns-668d6bf9bc-csrvc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" Nov 24 00:15:57.658836 containerd[1537]: 2025-11-24 00:15:57.601 [INFO][4614] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to califf78334c3e7 ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Namespace="kube-system" Pod="coredns-668d6bf9bc-csrvc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" Nov 24 00:15:57.658836 containerd[1537]: 2025-11-24 00:15:57.604 [INFO][4614] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Namespace="kube-system" Pod="coredns-668d6bf9bc-csrvc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" Nov 24 00:15:57.658966 containerd[1537]: 2025-11-24 00:15:57.604 [INFO][4614] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Namespace="kube-system" Pod="coredns-668d6bf9bc-csrvc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--csrvc-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"6978f440-5fc4-460d-8ab7-a825799d091f", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 14, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2", Pod:"coredns-668d6bf9bc-csrvc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"califf78334c3e7", MAC:"d6:a9:07:46:f0:ee", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:57.658966 containerd[1537]: 2025-11-24 00:15:57.645 [INFO][4614] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" Namespace="kube-system" Pod="coredns-668d6bf9bc-csrvc" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--csrvc-eth0" Nov 24 00:15:57.707537 containerd[1537]: time="2025-11-24T00:15:57.707476910Z" level=info msg="connecting to shim 34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2" address="unix:///run/containerd/s/b8c15d4c4b67cf4fed574ae71469a9e90756cd730ef17c2e1623af49af678264" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:57.753910 systemd[1]: Started cri-containerd-34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2.scope - libcontainer container 34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2. Nov 24 00:15:57.770273 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 24 00:15:57.870146 containerd[1537]: time="2025-11-24T00:15:57.869997070Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-csrvc,Uid:6978f440-5fc4-460d-8ab7-a825799d091f,Namespace:kube-system,Attempt:0,} returns sandbox id \"34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2\"" Nov 24 00:15:57.871067 kubelet[2683]: E1124 00:15:57.871022 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:57.883825 systemd-networkd[1448]: calic1e26248861: Link UP Nov 24 00:15:57.884710 systemd-networkd[1448]: calic1e26248861: Gained carrier Nov 24 00:15:57.887475 containerd[1537]: time="2025-11-24T00:15:57.887383659Z" level=info msg="CreateContainer within sandbox \"34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.650 [INFO][4636] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--666569f655--hr2wl-eth0 goldmane-666569f655- calico-system 22c38642-d71a-4cc4-996b-dcdaded5d62c 851 0 2025-11-24 00:15:12 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-666569f655-hr2wl eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calic1e26248861 [] [] }} ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Namespace="calico-system" Pod="goldmane-666569f655-hr2wl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--hr2wl-" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.650 [INFO][4636] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Namespace="calico-system" Pod="goldmane-666569f655-hr2wl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--hr2wl-eth0" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.720 [INFO][4676] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" HandleID="k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Workload="localhost-k8s-goldmane--666569f655--hr2wl-eth0" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.720 [INFO][4676] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" HandleID="k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Workload="localhost-k8s-goldmane--666569f655--hr2wl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e760), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-666569f655-hr2wl", "timestamp":"2025-11-24 00:15:57.720688761 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.721 [INFO][4676] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.721 [INFO][4676] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.721 [INFO][4676] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.729 [INFO][4676] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.737 [INFO][4676] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.742 [INFO][4676] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.748 [INFO][4676] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.750 [INFO][4676] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.750 [INFO][4676] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.752 [INFO][4676] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.756 [INFO][4676] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.874 [INFO][4676] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.874 [INFO][4676] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" host="localhost" Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.875 [INFO][4676] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:15:57.904039 containerd[1537]: 2025-11-24 00:15:57.875 [INFO][4676] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" HandleID="k8s-pod-network.497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Workload="localhost-k8s-goldmane--666569f655--hr2wl-eth0" Nov 24 00:15:57.905092 containerd[1537]: 2025-11-24 00:15:57.880 [INFO][4636] cni-plugin/k8s.go 418: Populated endpoint ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Namespace="calico-system" Pod="goldmane-666569f655-hr2wl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--hr2wl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--hr2wl-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"22c38642-d71a-4cc4-996b-dcdaded5d62c", ResourceVersion:"851", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-666569f655-hr2wl", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calic1e26248861", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:57.905092 containerd[1537]: 2025-11-24 00:15:57.880 [INFO][4636] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Namespace="calico-system" Pod="goldmane-666569f655-hr2wl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--hr2wl-eth0" Nov 24 00:15:57.905092 containerd[1537]: 2025-11-24 00:15:57.880 [INFO][4636] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic1e26248861 ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Namespace="calico-system" Pod="goldmane-666569f655-hr2wl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--hr2wl-eth0" Nov 24 00:15:57.905092 containerd[1537]: 2025-11-24 00:15:57.887 [INFO][4636] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Namespace="calico-system" Pod="goldmane-666569f655-hr2wl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--hr2wl-eth0" Nov 24 00:15:57.905092 containerd[1537]: 2025-11-24 00:15:57.889 [INFO][4636] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Namespace="calico-system" Pod="goldmane-666569f655-hr2wl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--hr2wl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--hr2wl-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"22c38642-d71a-4cc4-996b-dcdaded5d62c", ResourceVersion:"851", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b", Pod:"goldmane-666569f655-hr2wl", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calic1e26248861", MAC:"d6:05:c4:09:14:d8", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:57.905092 containerd[1537]: 2025-11-24 00:15:57.899 [INFO][4636] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" Namespace="calico-system" Pod="goldmane-666569f655-hr2wl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--hr2wl-eth0" Nov 24 00:15:57.914973 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1244077385.mount: Deactivated successfully. Nov 24 00:15:57.923418 containerd[1537]: time="2025-11-24T00:15:57.923308513Z" level=info msg="Container daa0fb659a971c5bf9ef873b55c22602144d94518bdf09d82704b8765864465a: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:15:57.934784 containerd[1537]: time="2025-11-24T00:15:57.934749682Z" level=info msg="CreateContainer within sandbox \"34b5a3db06caaa125d57e4e8be67efd01e5a0b505ccdecf2a08f29cb43a5b6a2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"daa0fb659a971c5bf9ef873b55c22602144d94518bdf09d82704b8765864465a\"" Nov 24 00:15:57.938095 containerd[1537]: time="2025-11-24T00:15:57.938064742Z" level=info msg="StartContainer for \"daa0fb659a971c5bf9ef873b55c22602144d94518bdf09d82704b8765864465a\"" Nov 24 00:15:57.939345 containerd[1537]: time="2025-11-24T00:15:57.939304998Z" level=info msg="connecting to shim daa0fb659a971c5bf9ef873b55c22602144d94518bdf09d82704b8765864465a" address="unix:///run/containerd/s/b8c15d4c4b67cf4fed574ae71469a9e90756cd730ef17c2e1623af49af678264" protocol=ttrpc version=3 Nov 24 00:15:57.952189 containerd[1537]: time="2025-11-24T00:15:57.952134457Z" level=info msg="connecting to shim 497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b" address="unix:///run/containerd/s/76223c6e98b7ca6df3a22657ce65f626a06db0ac805a7ba9f0c36312ee4473a3" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:57.960191 systemd-networkd[1448]: calif58a6519559: Link UP Nov 24 00:15:57.961955 systemd-networkd[1448]: calif58a6519559: Gained carrier Nov 24 00:15:57.967060 systemd[1]: Started cri-containerd-daa0fb659a971c5bf9ef873b55c22602144d94518bdf09d82704b8765864465a.scope - libcontainer container daa0fb659a971c5bf9ef873b55c22602144d94518bdf09d82704b8765864465a. Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.706 [INFO][4662] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0 calico-apiserver-c779694df- calico-apiserver 3b61fcb5-22ca-4d05-817b-ade781b1a990 848 0 2025-11-24 00:15:09 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:c779694df projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-c779694df-9w4qh eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calif58a6519559 [] [] }} ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-9w4qh" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--9w4qh-" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.706 [INFO][4662] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-9w4qh" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.746 [INFO][4701] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" HandleID="k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Workload="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.747 [INFO][4701] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" HandleID="k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Workload="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001384d0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-c779694df-9w4qh", "timestamp":"2025-11-24 00:15:57.746827196 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.747 [INFO][4701] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.875 [INFO][4701] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.875 [INFO][4701] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.885 [INFO][4701] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.897 [INFO][4701] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.906 [INFO][4701] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.909 [INFO][4701] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.914 [INFO][4701] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.918 [INFO][4701] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.924 [INFO][4701] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283 Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.930 [INFO][4701] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.941 [INFO][4701] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.941 [INFO][4701] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" host="localhost" Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.942 [INFO][4701] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:15:57.986001 containerd[1537]: 2025-11-24 00:15:57.942 [INFO][4701] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" HandleID="k8s-pod-network.48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Workload="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" Nov 24 00:15:57.986876 containerd[1537]: 2025-11-24 00:15:57.953 [INFO][4662] cni-plugin/k8s.go 418: Populated endpoint ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-9w4qh" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0", GenerateName:"calico-apiserver-c779694df-", Namespace:"calico-apiserver", SelfLink:"", UID:"3b61fcb5-22ca-4d05-817b-ade781b1a990", ResourceVersion:"848", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"c779694df", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-c779694df-9w4qh", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif58a6519559", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:57.986876 containerd[1537]: 2025-11-24 00:15:57.955 [INFO][4662] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-9w4qh" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" Nov 24 00:15:57.986876 containerd[1537]: 2025-11-24 00:15:57.955 [INFO][4662] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif58a6519559 ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-9w4qh" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" Nov 24 00:15:57.986876 containerd[1537]: 2025-11-24 00:15:57.962 [INFO][4662] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-9w4qh" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" Nov 24 00:15:57.986876 containerd[1537]: 2025-11-24 00:15:57.963 [INFO][4662] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-9w4qh" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0", GenerateName:"calico-apiserver-c779694df-", Namespace:"calico-apiserver", SelfLink:"", UID:"3b61fcb5-22ca-4d05-817b-ade781b1a990", ResourceVersion:"848", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"c779694df", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283", Pod:"calico-apiserver-c779694df-9w4qh", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif58a6519559", MAC:"0e:69:6e:bc:72:65", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:57.986876 containerd[1537]: 2025-11-24 00:15:57.982 [INFO][4662] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-9w4qh" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--9w4qh-eth0" Nov 24 00:15:57.991080 systemd[1]: Started cri-containerd-497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b.scope - libcontainer container 497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b. Nov 24 00:15:58.024347 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 24 00:15:58.026527 containerd[1537]: time="2025-11-24T00:15:58.026482182Z" level=info msg="connecting to shim 48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283" address="unix:///run/containerd/s/bc08d25f846e2dd2288abd5841f26d751023a867c5628f3874a4355548613021" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:58.064123 systemd[1]: Started cri-containerd-48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283.scope - libcontainer container 48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283. Nov 24 00:15:58.105805 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 24 00:15:58.192327 containerd[1537]: time="2025-11-24T00:15:58.190776549Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5jkmb,Uid:e7cdfd74-3050-4e07-b3f8-0110308ddcf6,Namespace:calico-system,Attempt:0,}" Nov 24 00:15:58.192327 containerd[1537]: time="2025-11-24T00:15:58.192120443Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fskrk,Uid:d27f4aa2-b9ea-448e-a394-15df266e8761,Namespace:kube-system,Attempt:0,}" Nov 24 00:15:58.192514 kubelet[2683]: E1124 00:15:58.190924 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:58.300131 containerd[1537]: time="2025-11-24T00:15:58.300076276Z" level=info msg="StartContainer for \"daa0fb659a971c5bf9ef873b55c22602144d94518bdf09d82704b8765864465a\" returns successfully" Nov 24 00:15:58.361374 containerd[1537]: time="2025-11-24T00:15:58.361294713Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-hr2wl,Uid:22c38642-d71a-4cc4-996b-dcdaded5d62c,Namespace:calico-system,Attempt:0,} returns sandbox id \"497d48cca77b5113bd88caae6e77187d6ef58f60c1462bdc37fce1362ee0ec3b\"" Nov 24 00:15:58.363031 containerd[1537]: time="2025-11-24T00:15:58.362984338Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 24 00:15:58.421276 containerd[1537]: time="2025-11-24T00:15:58.421227678Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-9w4qh,Uid:3b61fcb5-22ca-4d05-817b-ade781b1a990,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"48f2cf651c703185a265b45f74e66ef00ed5caf7711f4999a10f2be5ea593283\"" Nov 24 00:15:58.647042 kubelet[2683]: E1124 00:15:58.646761 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:58.705613 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3526254380.mount: Deactivated successfully. Nov 24 00:15:58.821868 systemd-networkd[1448]: caliadc74feb485: Link UP Nov 24 00:15:58.822972 systemd-networkd[1448]: caliadc74feb485: Gained carrier Nov 24 00:15:58.841709 kubelet[2683]: I1124 00:15:58.841500 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-csrvc" podStartSLOduration=67.841472205 podStartE2EDuration="1m7.841472205s" podCreationTimestamp="2025-11-24 00:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:15:58.668186883 +0000 UTC m=+73.575832265" watchObservedRunningTime="2025-11-24 00:15:58.841472205 +0000 UTC m=+73.749117577" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.727 [INFO][4883] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--5jkmb-eth0 csi-node-driver- calico-system e7cdfd74-3050-4e07-b3f8-0110308ddcf6 718 0 2025-11-24 00:15:14 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-5jkmb eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] caliadc74feb485 [] [] }} ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Namespace="calico-system" Pod="csi-node-driver-5jkmb" WorkloadEndpoint="localhost-k8s-csi--node--driver--5jkmb-" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.731 [INFO][4883] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Namespace="calico-system" Pod="csi-node-driver-5jkmb" WorkloadEndpoint="localhost-k8s-csi--node--driver--5jkmb-eth0" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.771 [INFO][4921] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" HandleID="k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Workload="localhost-k8s-csi--node--driver--5jkmb-eth0" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.771 [INFO][4921] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" HandleID="k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Workload="localhost-k8s-csi--node--driver--5jkmb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e760), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-5jkmb", "timestamp":"2025-11-24 00:15:58.771670154 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.772 [INFO][4921] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.772 [INFO][4921] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.772 [INFO][4921] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.782 [INFO][4921] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.787 [INFO][4921] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.794 [INFO][4921] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.797 [INFO][4921] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.800 [INFO][4921] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.800 [INFO][4921] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.802 [INFO][4921] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6 Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.807 [INFO][4921] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.814 [INFO][4921] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.814 [INFO][4921] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" host="localhost" Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.814 [INFO][4921] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:15:58.843644 containerd[1537]: 2025-11-24 00:15:58.814 [INFO][4921] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" HandleID="k8s-pod-network.27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Workload="localhost-k8s-csi--node--driver--5jkmb-eth0" Nov 24 00:15:58.844606 containerd[1537]: 2025-11-24 00:15:58.817 [INFO][4883] cni-plugin/k8s.go 418: Populated endpoint ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Namespace="calico-system" Pod="csi-node-driver-5jkmb" WorkloadEndpoint="localhost-k8s-csi--node--driver--5jkmb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--5jkmb-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"e7cdfd74-3050-4e07-b3f8-0110308ddcf6", ResourceVersion:"718", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-5jkmb", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"caliadc74feb485", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:58.844606 containerd[1537]: 2025-11-24 00:15:58.818 [INFO][4883] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Namespace="calico-system" Pod="csi-node-driver-5jkmb" WorkloadEndpoint="localhost-k8s-csi--node--driver--5jkmb-eth0" Nov 24 00:15:58.844606 containerd[1537]: 2025-11-24 00:15:58.818 [INFO][4883] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliadc74feb485 ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Namespace="calico-system" Pod="csi-node-driver-5jkmb" WorkloadEndpoint="localhost-k8s-csi--node--driver--5jkmb-eth0" Nov 24 00:15:58.844606 containerd[1537]: 2025-11-24 00:15:58.823 [INFO][4883] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Namespace="calico-system" Pod="csi-node-driver-5jkmb" WorkloadEndpoint="localhost-k8s-csi--node--driver--5jkmb-eth0" Nov 24 00:15:58.844606 containerd[1537]: 2025-11-24 00:15:58.823 [INFO][4883] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Namespace="calico-system" Pod="csi-node-driver-5jkmb" WorkloadEndpoint="localhost-k8s-csi--node--driver--5jkmb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--5jkmb-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"e7cdfd74-3050-4e07-b3f8-0110308ddcf6", ResourceVersion:"718", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6", Pod:"csi-node-driver-5jkmb", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"caliadc74feb485", MAC:"ca:62:b4:c0:5c:3a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:58.844606 containerd[1537]: 2025-11-24 00:15:58.840 [INFO][4883] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" Namespace="calico-system" Pod="csi-node-driver-5jkmb" WorkloadEndpoint="localhost-k8s-csi--node--driver--5jkmb-eth0" Nov 24 00:15:58.884003 containerd[1537]: time="2025-11-24T00:15:58.883875515Z" level=info msg="connecting to shim 27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6" address="unix:///run/containerd/s/b8e3a2ef29516a9c5b4f1ddee717e384c945bed1e107e2f8f320248ae3b0216c" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:58.926919 systemd[1]: Started cri-containerd-27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6.scope - libcontainer container 27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6. Nov 24 00:15:58.945791 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 24 00:15:58.947329 systemd-networkd[1448]: cali4f1019ee5b1: Link UP Nov 24 00:15:58.948344 systemd-networkd[1448]: cali4f1019ee5b1: Gained carrier Nov 24 00:15:58.966752 containerd[1537]: time="2025-11-24T00:15:58.966711293Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:15:59.095086 containerd[1537]: time="2025-11-24T00:15:59.095037093Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5jkmb,Uid:e7cdfd74-3050-4e07-b3f8-0110308ddcf6,Namespace:calico-system,Attempt:0,} returns sandbox id \"27638f3a5d3309dba6d3e1843f7bc967bc6ebfb0063742a637daeebd397547c6\"" Nov 24 00:15:59.132849 systemd-networkd[1448]: calic1e26248861: Gained IPv6LL Nov 24 00:15:59.190910 containerd[1537]: time="2025-11-24T00:15:59.190747885Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-n2448,Uid:6cfea110-fbd7-44cd-904d-4c7d268ca542,Namespace:calico-apiserver,Attempt:0,}" Nov 24 00:15:59.225330 containerd[1537]: time="2025-11-24T00:15:59.225138128Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 24 00:15:59.225330 containerd[1537]: time="2025-11-24T00:15:59.225207329Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 24 00:15:59.226107 kubelet[2683]: E1124 00:15:59.226007 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:15:59.226107 kubelet[2683]: E1124 00:15:59.226098 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:15:59.226571 kubelet[2683]: E1124 00:15:59.226425 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t59pv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-hr2wl_calico-system(22c38642-d71a-4cc4-996b-dcdaded5d62c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 24 00:15:59.226774 containerd[1537]: time="2025-11-24T00:15:59.226528217Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:15:59.228093 kubelet[2683]: E1124 00:15:59.227890 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-hr2wl" podUID="22c38642-d71a-4cc4-996b-dcdaded5d62c" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.717 [INFO][4894] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--fskrk-eth0 coredns-668d6bf9bc- kube-system d27f4aa2-b9ea-448e-a394-15df266e8761 845 0 2025-11-24 00:14:51 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-fskrk eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali4f1019ee5b1 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Namespace="kube-system" Pod="coredns-668d6bf9bc-fskrk" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--fskrk-" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.718 [INFO][4894] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Namespace="kube-system" Pod="coredns-668d6bf9bc-fskrk" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.774 [INFO][4915] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" HandleID="k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Workload="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.775 [INFO][4915] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" HandleID="k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Workload="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002c75e0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-fskrk", "timestamp":"2025-11-24 00:15:58.774829293 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.775 [INFO][4915] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.814 [INFO][4915] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.814 [INFO][4915] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.883 [INFO][4915] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.890 [INFO][4915] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.902 [INFO][4915] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.908 [INFO][4915] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.915 [INFO][4915] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.915 [INFO][4915] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.917 [INFO][4915] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91 Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.926 [INFO][4915] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.938 [INFO][4915] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.938 [INFO][4915] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" host="localhost" Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.939 [INFO][4915] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:15:59.240861 containerd[1537]: 2025-11-24 00:15:58.939 [INFO][4915] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" HandleID="k8s-pod-network.e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Workload="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" Nov 24 00:15:59.241600 containerd[1537]: 2025-11-24 00:15:58.943 [INFO][4894] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Namespace="kube-system" Pod="coredns-668d6bf9bc-fskrk" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--fskrk-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"d27f4aa2-b9ea-448e-a394-15df266e8761", ResourceVersion:"845", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 14, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-fskrk", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4f1019ee5b1", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:59.241600 containerd[1537]: 2025-11-24 00:15:58.943 [INFO][4894] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Namespace="kube-system" Pod="coredns-668d6bf9bc-fskrk" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" Nov 24 00:15:59.241600 containerd[1537]: 2025-11-24 00:15:58.944 [INFO][4894] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4f1019ee5b1 ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Namespace="kube-system" Pod="coredns-668d6bf9bc-fskrk" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" Nov 24 00:15:59.241600 containerd[1537]: 2025-11-24 00:15:58.948 [INFO][4894] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Namespace="kube-system" Pod="coredns-668d6bf9bc-fskrk" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" Nov 24 00:15:59.241600 containerd[1537]: 2025-11-24 00:15:58.949 [INFO][4894] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Namespace="kube-system" Pod="coredns-668d6bf9bc-fskrk" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--fskrk-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"d27f4aa2-b9ea-448e-a394-15df266e8761", ResourceVersion:"845", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 14, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91", Pod:"coredns-668d6bf9bc-fskrk", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4f1019ee5b1", MAC:"12:ca:5b:b5:eb:09", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:59.241600 containerd[1537]: 2025-11-24 00:15:59.234 [INFO][4894] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" Namespace="kube-system" Pod="coredns-668d6bf9bc-fskrk" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--fskrk-eth0" Nov 24 00:15:59.275736 containerd[1537]: time="2025-11-24T00:15:59.275148572Z" level=info msg="connecting to shim e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91" address="unix:///run/containerd/s/38afc8424e5930c2456fb126bf9c499b5852540c6228bff6d7f82380bfb6d204" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:59.304936 systemd[1]: Started cri-containerd-e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91.scope - libcontainer container e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91. Nov 24 00:15:59.325928 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 24 00:15:59.370200 systemd-networkd[1448]: cali89392157d93: Link UP Nov 24 00:15:59.371161 systemd-networkd[1448]: cali89392157d93: Gained carrier Nov 24 00:15:59.371752 containerd[1537]: time="2025-11-24T00:15:59.371320266Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fskrk,Uid:d27f4aa2-b9ea-448e-a394-15df266e8761,Namespace:kube-system,Attempt:0,} returns sandbox id \"e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91\"" Nov 24 00:15:59.372516 kubelet[2683]: E1124 00:15:59.372412 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:59.374848 containerd[1537]: time="2025-11-24T00:15:59.374812789Z" level=info msg="CreateContainer within sandbox \"e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 24 00:15:59.395270 containerd[1537]: time="2025-11-24T00:15:59.394672976Z" level=info msg="Container a32a6dd40d023d666e9bcc167078674b10bfd306531e394229bc01b06dca1a3c: CDI devices from CRI Config.CDIDevices: []" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.275 [INFO][4992] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--c779694df--n2448-eth0 calico-apiserver-c779694df- calico-apiserver 6cfea110-fbd7-44cd-904d-4c7d268ca542 853 0 2025-11-24 00:15:09 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:c779694df projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-c779694df-n2448 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali89392157d93 [] [] }} ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-n2448" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--n2448-" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.275 [INFO][4992] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-n2448" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.308 [INFO][5037] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" HandleID="k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Workload="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.308 [INFO][5037] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" HandleID="k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Workload="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002c6950), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-c779694df-n2448", "timestamp":"2025-11-24 00:15:59.308196175 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.308 [INFO][5037] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.308 [INFO][5037] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.308 [INFO][5037] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.318 [INFO][5037] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.326 [INFO][5037] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.333 [INFO][5037] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.335 [INFO][5037] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.338 [INFO][5037] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.338 [INFO][5037] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.340 [INFO][5037] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9 Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.344 [INFO][5037] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.357 [INFO][5037] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.357 [INFO][5037] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" host="localhost" Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.357 [INFO][5037] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 24 00:15:59.397777 containerd[1537]: 2025-11-24 00:15:59.357 [INFO][5037] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" HandleID="k8s-pod-network.c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Workload="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" Nov 24 00:15:59.398616 containerd[1537]: 2025-11-24 00:15:59.363 [INFO][4992] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-n2448" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--c779694df--n2448-eth0", GenerateName:"calico-apiserver-c779694df-", Namespace:"calico-apiserver", SelfLink:"", UID:"6cfea110-fbd7-44cd-904d-4c7d268ca542", ResourceVersion:"853", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"c779694df", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-c779694df-n2448", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali89392157d93", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:59.398616 containerd[1537]: 2025-11-24 00:15:59.364 [INFO][4992] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-n2448" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" Nov 24 00:15:59.398616 containerd[1537]: 2025-11-24 00:15:59.364 [INFO][4992] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali89392157d93 ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-n2448" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" Nov 24 00:15:59.398616 containerd[1537]: 2025-11-24 00:15:59.372 [INFO][4992] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-n2448" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" Nov 24 00:15:59.398616 containerd[1537]: 2025-11-24 00:15:59.372 [INFO][4992] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-n2448" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--c779694df--n2448-eth0", GenerateName:"calico-apiserver-c779694df-", Namespace:"calico-apiserver", SelfLink:"", UID:"6cfea110-fbd7-44cd-904d-4c7d268ca542", ResourceVersion:"853", Generation:0, CreationTimestamp:time.Date(2025, time.November, 24, 0, 15, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"c779694df", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9", Pod:"calico-apiserver-c779694df-n2448", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali89392157d93", MAC:"b2:e2:b5:06:28:54", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 24 00:15:59.398616 containerd[1537]: 2025-11-24 00:15:59.390 [INFO][4992] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" Namespace="calico-apiserver" Pod="calico-apiserver-c779694df-n2448" WorkloadEndpoint="localhost-k8s-calico--apiserver--c779694df--n2448-eth0" Nov 24 00:15:59.405749 containerd[1537]: time="2025-11-24T00:15:59.405704487Z" level=info msg="CreateContainer within sandbox \"e51602a2387d2469b1e616788fa5cc517fb57c03fa9e16cc7a8f55e3f0fb4b91\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a32a6dd40d023d666e9bcc167078674b10bfd306531e394229bc01b06dca1a3c\"" Nov 24 00:15:59.407007 containerd[1537]: time="2025-11-24T00:15:59.406883623Z" level=info msg="StartContainer for \"a32a6dd40d023d666e9bcc167078674b10bfd306531e394229bc01b06dca1a3c\"" Nov 24 00:15:59.407885 containerd[1537]: time="2025-11-24T00:15:59.407850844Z" level=info msg="connecting to shim a32a6dd40d023d666e9bcc167078674b10bfd306531e394229bc01b06dca1a3c" address="unix:///run/containerd/s/38afc8424e5930c2456fb126bf9c499b5852540c6228bff6d7f82380bfb6d204" protocol=ttrpc version=3 Nov 24 00:15:59.430880 systemd[1]: Started cri-containerd-a32a6dd40d023d666e9bcc167078674b10bfd306531e394229bc01b06dca1a3c.scope - libcontainer container a32a6dd40d023d666e9bcc167078674b10bfd306531e394229bc01b06dca1a3c. Nov 24 00:15:59.437540 containerd[1537]: time="2025-11-24T00:15:59.437413218Z" level=info msg="connecting to shim c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9" address="unix:///run/containerd/s/248cf3ebd331ee87cb1ca6390f37e9992823e74e42940740641a42fe97234187" namespace=k8s.io protocol=ttrpc version=3 Nov 24 00:15:59.464817 systemd[1]: Started cri-containerd-c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9.scope - libcontainer container c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9. Nov 24 00:15:59.485438 containerd[1537]: time="2025-11-24T00:15:59.485382776Z" level=info msg="StartContainer for \"a32a6dd40d023d666e9bcc167078674b10bfd306531e394229bc01b06dca1a3c\" returns successfully" Nov 24 00:15:59.491085 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 24 00:15:59.517393 systemd-networkd[1448]: califf78334c3e7: Gained IPv6LL Nov 24 00:15:59.530215 containerd[1537]: time="2025-11-24T00:15:59.530084843Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c779694df-n2448,Uid:6cfea110-fbd7-44cd-904d-4c7d268ca542,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"c928ee9fc41a0c4102f2602d8fc85639f8419ccdd5c7b21962bfddbdc4a423d9\"" Nov 24 00:15:59.554700 containerd[1537]: time="2025-11-24T00:15:59.554533460Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:15:59.556255 containerd[1537]: time="2025-11-24T00:15:59.556190802Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:15:59.556255 containerd[1537]: time="2025-11-24T00:15:59.556250997Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:15:59.556510 kubelet[2683]: E1124 00:15:59.556462 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:15:59.556652 kubelet[2683]: E1124 00:15:59.556514 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:15:59.556840 kubelet[2683]: E1124 00:15:59.556785 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rbpm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-c779694df-9w4qh_calico-apiserver(3b61fcb5-22ca-4d05-817b-ade781b1a990): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:15:59.557098 containerd[1537]: time="2025-11-24T00:15:59.556860173Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 24 00:15:59.559044 kubelet[2683]: E1124 00:15:59.558890 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" podUID="3b61fcb5-22ca-4d05-817b-ade781b1a990" Nov 24 00:15:59.653537 kubelet[2683]: E1124 00:15:59.653267 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:59.655283 kubelet[2683]: E1124 00:15:59.655249 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:15:59.655604 kubelet[2683]: E1124 00:15:59.655579 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" podUID="3b61fcb5-22ca-4d05-817b-ade781b1a990" Nov 24 00:15:59.656123 kubelet[2683]: E1124 00:15:59.656083 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-hr2wl" podUID="22c38642-d71a-4cc4-996b-dcdaded5d62c" Nov 24 00:15:59.772893 systemd-networkd[1448]: calif58a6519559: Gained IPv6LL Nov 24 00:15:59.877538 containerd[1537]: time="2025-11-24T00:15:59.877454042Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:15:59.901816 systemd-networkd[1448]: caliadc74feb485: Gained IPv6LL Nov 24 00:15:59.921508 containerd[1537]: time="2025-11-24T00:15:59.921417926Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 24 00:15:59.921508 containerd[1537]: time="2025-11-24T00:15:59.921494803Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 24 00:15:59.921879 kubelet[2683]: E1124 00:15:59.921829 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:15:59.921990 kubelet[2683]: E1124 00:15:59.921895 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:15:59.922235 kubelet[2683]: E1124 00:15:59.922186 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fkmrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 24 00:15:59.922447 containerd[1537]: time="2025-11-24T00:15:59.922358676Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:16:00.385720 containerd[1537]: time="2025-11-24T00:16:00.385653611Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:00.497313 containerd[1537]: time="2025-11-24T00:16:00.497209526Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:16:00.497501 containerd[1537]: time="2025-11-24T00:16:00.497286253Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:16:00.497697 kubelet[2683]: E1124 00:16:00.497582 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:00.497697 kubelet[2683]: E1124 00:16:00.497686 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:00.498076 kubelet[2683]: E1124 00:16:00.497981 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hbz2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-c779694df-n2448_calico-apiserver(6cfea110-fbd7-44cd-904d-4c7d268ca542): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:00.498182 containerd[1537]: time="2025-11-24T00:16:00.498115599Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 24 00:16:00.499312 kubelet[2683]: E1124 00:16:00.499246 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-n2448" podUID="6cfea110-fbd7-44cd-904d-4c7d268ca542" Nov 24 00:16:00.540812 systemd-networkd[1448]: cali4f1019ee5b1: Gained IPv6LL Nov 24 00:16:00.605658 kubelet[2683]: I1124 00:16:00.603860 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-fskrk" podStartSLOduration=69.603831392 podStartE2EDuration="1m9.603831392s" podCreationTimestamp="2025-11-24 00:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 00:16:00.128487928 +0000 UTC m=+75.036133290" watchObservedRunningTime="2025-11-24 00:16:00.603831392 +0000 UTC m=+75.511476744" Nov 24 00:16:00.656873 kubelet[2683]: E1124 00:16:00.656718 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:00.657065 kubelet[2683]: E1124 00:16:00.656925 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:00.657544 kubelet[2683]: E1124 00:16:00.657507 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-n2448" podUID="6cfea110-fbd7-44cd-904d-4c7d268ca542" Nov 24 00:16:00.845113 systemd[1]: Started sshd@10-10.0.0.25:22-10.0.0.1:54878.service - OpenSSH per-connection server daemon (10.0.0.1:54878). Nov 24 00:16:00.925375 sshd[5172]: Accepted publickey for core from 10.0.0.1 port 54878 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:00.927270 sshd-session[5172]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:00.932512 systemd-logind[1519]: New session 11 of user core. Nov 24 00:16:00.937771 systemd[1]: Started session-11.scope - Session 11 of User core. Nov 24 00:16:01.074105 containerd[1537]: time="2025-11-24T00:16:01.073905056Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:01.076363 containerd[1537]: time="2025-11-24T00:16:01.076308932Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 24 00:16:01.076586 containerd[1537]: time="2025-11-24T00:16:01.076466905Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 24 00:16:01.077230 kubelet[2683]: E1124 00:16:01.077182 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:16:01.077733 kubelet[2683]: E1124 00:16:01.077234 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:16:01.077733 kubelet[2683]: E1124 00:16:01.077338 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fkmrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:01.079704 kubelet[2683]: E1124 00:16:01.079588 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:16:01.109699 sshd[5177]: Connection closed by 10.0.0.1 port 54878 Nov 24 00:16:01.110108 sshd-session[5172]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:01.118758 systemd[1]: sshd@10-10.0.0.25:22-10.0.0.1:54878.service: Deactivated successfully. Nov 24 00:16:01.120811 systemd[1]: session-11.scope: Deactivated successfully. Nov 24 00:16:01.121716 systemd-logind[1519]: Session 11 logged out. Waiting for processes to exit. Nov 24 00:16:01.124807 systemd[1]: Started sshd@11-10.0.0.25:22-10.0.0.1:54880.service - OpenSSH per-connection server daemon (10.0.0.1:54880). Nov 24 00:16:01.125956 systemd-logind[1519]: Removed session 11. Nov 24 00:16:01.173656 sshd[5191]: Accepted publickey for core from 10.0.0.1 port 54880 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:01.175444 sshd-session[5191]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:01.181055 systemd-logind[1519]: New session 12 of user core. Nov 24 00:16:01.187832 systemd[1]: Started session-12.scope - Session 12 of User core. Nov 24 00:16:01.372872 systemd-networkd[1448]: cali89392157d93: Gained IPv6LL Nov 24 00:16:01.534266 sshd[5194]: Connection closed by 10.0.0.1 port 54880 Nov 24 00:16:01.534757 sshd-session[5191]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:01.543662 systemd[1]: sshd@11-10.0.0.25:22-10.0.0.1:54880.service: Deactivated successfully. Nov 24 00:16:01.545918 systemd[1]: session-12.scope: Deactivated successfully. Nov 24 00:16:01.546856 systemd-logind[1519]: Session 12 logged out. Waiting for processes to exit. Nov 24 00:16:01.550441 systemd[1]: Started sshd@12-10.0.0.25:22-10.0.0.1:54894.service - OpenSSH per-connection server daemon (10.0.0.1:54894). Nov 24 00:16:01.551142 systemd-logind[1519]: Removed session 12. Nov 24 00:16:01.601340 sshd[5205]: Accepted publickey for core from 10.0.0.1 port 54894 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:01.603469 sshd-session[5205]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:01.609693 systemd-logind[1519]: New session 13 of user core. Nov 24 00:16:01.614903 systemd[1]: Started session-13.scope - Session 13 of User core. Nov 24 00:16:01.658703 kubelet[2683]: E1124 00:16:01.658651 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:01.660699 kubelet[2683]: E1124 00:16:01.660597 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:16:01.863719 sshd[5208]: Connection closed by 10.0.0.1 port 54894 Nov 24 00:16:01.864015 sshd-session[5205]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:01.869445 systemd[1]: sshd@12-10.0.0.25:22-10.0.0.1:54894.service: Deactivated successfully. Nov 24 00:16:01.871561 systemd[1]: session-13.scope: Deactivated successfully. Nov 24 00:16:01.872430 systemd-logind[1519]: Session 13 logged out. Waiting for processes to exit. Nov 24 00:16:01.874026 systemd-logind[1519]: Removed session 13. Nov 24 00:16:04.192011 containerd[1537]: time="2025-11-24T00:16:04.191751680Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 24 00:16:04.579847 containerd[1537]: time="2025-11-24T00:16:04.579785932Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:04.672112 containerd[1537]: time="2025-11-24T00:16:04.672006847Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 24 00:16:04.672112 containerd[1537]: time="2025-11-24T00:16:04.672117198Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 24 00:16:04.672357 kubelet[2683]: E1124 00:16:04.672236 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:16:04.672357 kubelet[2683]: E1124 00:16:04.672290 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:16:04.672889 kubelet[2683]: E1124 00:16:04.672451 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:ed3f3941752e478da85a0af9ae40cdb3,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4wwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-7657d947d7-4pbrs_calico-system(4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:04.674734 containerd[1537]: time="2025-11-24T00:16:04.674673669Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 24 00:16:05.046261 containerd[1537]: time="2025-11-24T00:16:05.046189808Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:05.087585 containerd[1537]: time="2025-11-24T00:16:05.087388124Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 24 00:16:05.087585 containerd[1537]: time="2025-11-24T00:16:05.087451595Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 24 00:16:05.087831 kubelet[2683]: E1124 00:16:05.087783 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:16:05.087874 kubelet[2683]: E1124 00:16:05.087845 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:16:05.088069 kubelet[2683]: E1124 00:16:05.087995 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-7657d947d7-4pbrs_calico-system(4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:05.093106 kubelet[2683]: E1124 00:16:05.092384 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-7657d947d7-4pbrs" podUID="4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2" Nov 24 00:16:06.882659 systemd[1]: Started sshd@13-10.0.0.25:22-10.0.0.1:54906.service - OpenSSH per-connection server daemon (10.0.0.1:54906). Nov 24 00:16:06.945990 sshd[5226]: Accepted publickey for core from 10.0.0.1 port 54906 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:06.947843 sshd-session[5226]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:06.952714 systemd-logind[1519]: New session 14 of user core. Nov 24 00:16:06.961755 systemd[1]: Started session-14.scope - Session 14 of User core. Nov 24 00:16:07.082034 sshd[5229]: Connection closed by 10.0.0.1 port 54906 Nov 24 00:16:07.082465 sshd-session[5226]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:07.087439 systemd[1]: sshd@13-10.0.0.25:22-10.0.0.1:54906.service: Deactivated successfully. Nov 24 00:16:07.090090 systemd[1]: session-14.scope: Deactivated successfully. Nov 24 00:16:07.091435 systemd-logind[1519]: Session 14 logged out. Waiting for processes to exit. Nov 24 00:16:07.093296 systemd-logind[1519]: Removed session 14. Nov 24 00:16:10.190580 kubelet[2683]: E1124 00:16:10.190521 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:11.190318 kubelet[2683]: E1124 00:16:11.190258 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:12.097707 systemd[1]: Started sshd@14-10.0.0.25:22-10.0.0.1:52308.service - OpenSSH per-connection server daemon (10.0.0.1:52308). Nov 24 00:16:12.167894 sshd[5254]: Accepted publickey for core from 10.0.0.1 port 52308 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:12.170176 sshd-session[5254]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:12.177071 systemd-logind[1519]: New session 15 of user core. Nov 24 00:16:12.185913 systemd[1]: Started session-15.scope - Session 15 of User core. Nov 24 00:16:12.191332 containerd[1537]: time="2025-11-24T00:16:12.191289239Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 24 00:16:12.356658 sshd[5257]: Connection closed by 10.0.0.1 port 52308 Nov 24 00:16:12.357097 sshd-session[5254]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:12.361767 systemd-logind[1519]: Session 15 logged out. Waiting for processes to exit. Nov 24 00:16:12.364230 systemd[1]: sshd@14-10.0.0.25:22-10.0.0.1:52308.service: Deactivated successfully. Nov 24 00:16:12.368208 systemd[1]: session-15.scope: Deactivated successfully. Nov 24 00:16:12.371257 systemd-logind[1519]: Removed session 15. Nov 24 00:16:12.509584 containerd[1537]: time="2025-11-24T00:16:12.509514509Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:12.512868 containerd[1537]: time="2025-11-24T00:16:12.512815961Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 24 00:16:12.512868 containerd[1537]: time="2025-11-24T00:16:12.512909890Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 24 00:16:12.513205 kubelet[2683]: E1124 00:16:12.513119 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:16:12.513670 kubelet[2683]: E1124 00:16:12.513236 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:16:12.514602 kubelet[2683]: E1124 00:16:12.514070 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xn2qq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5889ff9675-8zv2t_calico-system(024a87f6-fa17-474a-b3f9-39d35fa916e3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:12.515414 kubelet[2683]: E1124 00:16:12.515259 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" podUID="024a87f6-fa17-474a-b3f9-39d35fa916e3" Nov 24 00:16:13.194526 containerd[1537]: time="2025-11-24T00:16:13.194473054Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 24 00:16:13.545314 containerd[1537]: time="2025-11-24T00:16:13.545232747Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:13.546477 containerd[1537]: time="2025-11-24T00:16:13.546442969Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 24 00:16:13.546649 containerd[1537]: time="2025-11-24T00:16:13.546557377Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 24 00:16:13.546783 kubelet[2683]: E1124 00:16:13.546739 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:16:13.547215 kubelet[2683]: E1124 00:16:13.546815 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:16:13.547549 kubelet[2683]: E1124 00:16:13.547387 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fkmrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:13.547678 containerd[1537]: time="2025-11-24T00:16:13.547607996Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:16:13.878259 containerd[1537]: time="2025-11-24T00:16:13.878075527Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:13.992084 containerd[1537]: time="2025-11-24T00:16:13.991966169Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:16:13.992084 containerd[1537]: time="2025-11-24T00:16:13.992024590Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:16:13.992499 kubelet[2683]: E1124 00:16:13.992426 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:13.992565 kubelet[2683]: E1124 00:16:13.992505 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:13.992965 kubelet[2683]: E1124 00:16:13.992874 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rbpm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-c779694df-9w4qh_calico-apiserver(3b61fcb5-22ca-4d05-817b-ade781b1a990): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:13.993408 containerd[1537]: time="2025-11-24T00:16:13.992978895Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 24 00:16:13.994383 kubelet[2683]: E1124 00:16:13.994306 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" podUID="3b61fcb5-22ca-4d05-817b-ade781b1a990" Nov 24 00:16:14.387851 containerd[1537]: time="2025-11-24T00:16:14.387794487Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:14.408246 containerd[1537]: time="2025-11-24T00:16:14.408135239Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 24 00:16:14.408450 containerd[1537]: time="2025-11-24T00:16:14.408269004Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 24 00:16:14.408545 kubelet[2683]: E1124 00:16:14.408486 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:16:14.408595 kubelet[2683]: E1124 00:16:14.408550 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:16:14.409065 containerd[1537]: time="2025-11-24T00:16:14.409030351Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:16:14.409160 kubelet[2683]: E1124 00:16:14.409057 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t59pv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-hr2wl_calico-system(22c38642-d71a-4cc4-996b-dcdaded5d62c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:14.410647 kubelet[2683]: E1124 00:16:14.410558 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-hr2wl" podUID="22c38642-d71a-4cc4-996b-dcdaded5d62c" Nov 24 00:16:14.808702 containerd[1537]: time="2025-11-24T00:16:14.808606790Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:14.810016 containerd[1537]: time="2025-11-24T00:16:14.809965795Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:16:14.810121 containerd[1537]: time="2025-11-24T00:16:14.810065194Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:16:14.810264 kubelet[2683]: E1124 00:16:14.810213 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:14.810732 kubelet[2683]: E1124 00:16:14.810273 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:14.810804 kubelet[2683]: E1124 00:16:14.810757 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hbz2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-c779694df-n2448_calico-apiserver(6cfea110-fbd7-44cd-904d-4c7d268ca542): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:14.811067 containerd[1537]: time="2025-11-24T00:16:14.810981096Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 24 00:16:14.812041 kubelet[2683]: E1124 00:16:14.811941 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-n2448" podUID="6cfea110-fbd7-44cd-904d-4c7d268ca542" Nov 24 00:16:15.115995 containerd[1537]: time="2025-11-24T00:16:15.115820910Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:15.119773 containerd[1537]: time="2025-11-24T00:16:15.119727317Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 24 00:16:15.119840 containerd[1537]: time="2025-11-24T00:16:15.119818710Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 24 00:16:15.120042 kubelet[2683]: E1124 00:16:15.119997 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:16:15.120133 kubelet[2683]: E1124 00:16:15.120054 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:16:15.120206 kubelet[2683]: E1124 00:16:15.120171 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fkmrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:15.121431 kubelet[2683]: E1124 00:16:15.121378 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:16:17.377010 systemd[1]: Started sshd@15-10.0.0.25:22-10.0.0.1:52318.service - OpenSSH per-connection server daemon (10.0.0.1:52318). Nov 24 00:16:17.458138 sshd[5270]: Accepted publickey for core from 10.0.0.1 port 52318 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:17.460078 sshd-session[5270]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:17.466329 systemd-logind[1519]: New session 16 of user core. Nov 24 00:16:17.475955 systemd[1]: Started session-16.scope - Session 16 of User core. Nov 24 00:16:17.608463 sshd[5273]: Connection closed by 10.0.0.1 port 52318 Nov 24 00:16:17.608839 sshd-session[5270]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:17.613674 systemd[1]: sshd@15-10.0.0.25:22-10.0.0.1:52318.service: Deactivated successfully. Nov 24 00:16:17.615835 systemd[1]: session-16.scope: Deactivated successfully. Nov 24 00:16:17.616730 systemd-logind[1519]: Session 16 logged out. Waiting for processes to exit. Nov 24 00:16:17.618670 systemd-logind[1519]: Removed session 16. Nov 24 00:16:17.724612 kubelet[2683]: E1124 00:16:17.724479 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:20.194080 kubelet[2683]: E1124 00:16:20.194019 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-7657d947d7-4pbrs" podUID="4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2" Nov 24 00:16:22.626037 systemd[1]: Started sshd@16-10.0.0.25:22-10.0.0.1:46750.service - OpenSSH per-connection server daemon (10.0.0.1:46750). Nov 24 00:16:22.686092 sshd[5316]: Accepted publickey for core from 10.0.0.1 port 46750 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:22.687766 sshd-session[5316]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:22.692778 systemd-logind[1519]: New session 17 of user core. Nov 24 00:16:22.703802 systemd[1]: Started session-17.scope - Session 17 of User core. Nov 24 00:16:22.830250 sshd[5319]: Connection closed by 10.0.0.1 port 46750 Nov 24 00:16:22.830752 sshd-session[5316]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:22.841444 systemd[1]: sshd@16-10.0.0.25:22-10.0.0.1:46750.service: Deactivated successfully. Nov 24 00:16:22.843571 systemd[1]: session-17.scope: Deactivated successfully. Nov 24 00:16:22.844583 systemd-logind[1519]: Session 17 logged out. Waiting for processes to exit. Nov 24 00:16:22.847523 systemd[1]: Started sshd@17-10.0.0.25:22-10.0.0.1:46756.service - OpenSSH per-connection server daemon (10.0.0.1:46756). Nov 24 00:16:22.848841 systemd-logind[1519]: Removed session 17. Nov 24 00:16:22.906847 sshd[5332]: Accepted publickey for core from 10.0.0.1 port 46756 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:22.908897 sshd-session[5332]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:22.913834 systemd-logind[1519]: New session 18 of user core. Nov 24 00:16:22.930814 systemd[1]: Started session-18.scope - Session 18 of User core. Nov 24 00:16:23.207944 kubelet[2683]: E1124 00:16:23.207733 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:23.264137 sshd[5335]: Connection closed by 10.0.0.1 port 46756 Nov 24 00:16:23.264992 sshd-session[5332]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:23.278955 systemd[1]: sshd@17-10.0.0.25:22-10.0.0.1:46756.service: Deactivated successfully. Nov 24 00:16:23.281229 systemd[1]: session-18.scope: Deactivated successfully. Nov 24 00:16:23.282294 systemd-logind[1519]: Session 18 logged out. Waiting for processes to exit. Nov 24 00:16:23.286306 systemd[1]: Started sshd@18-10.0.0.25:22-10.0.0.1:46768.service - OpenSSH per-connection server daemon (10.0.0.1:46768). Nov 24 00:16:23.286972 systemd-logind[1519]: Removed session 18. Nov 24 00:16:23.363695 sshd[5346]: Accepted publickey for core from 10.0.0.1 port 46768 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:23.365706 sshd-session[5346]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:23.370606 systemd-logind[1519]: New session 19 of user core. Nov 24 00:16:23.377849 systemd[1]: Started session-19.scope - Session 19 of User core. Nov 24 00:16:23.898073 sshd[5349]: Connection closed by 10.0.0.1 port 46768 Nov 24 00:16:23.898914 sshd-session[5346]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:23.909051 systemd[1]: sshd@18-10.0.0.25:22-10.0.0.1:46768.service: Deactivated successfully. Nov 24 00:16:23.912008 systemd[1]: session-19.scope: Deactivated successfully. Nov 24 00:16:23.916473 systemd-logind[1519]: Session 19 logged out. Waiting for processes to exit. Nov 24 00:16:23.928424 systemd[1]: Started sshd@19-10.0.0.25:22-10.0.0.1:46782.service - OpenSSH per-connection server daemon (10.0.0.1:46782). Nov 24 00:16:23.929751 systemd-logind[1519]: Removed session 19. Nov 24 00:16:23.979176 sshd[5367]: Accepted publickey for core from 10.0.0.1 port 46782 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:23.980647 sshd-session[5367]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:23.985104 systemd-logind[1519]: New session 20 of user core. Nov 24 00:16:23.996789 systemd[1]: Started session-20.scope - Session 20 of User core. Nov 24 00:16:24.235153 sshd[5370]: Connection closed by 10.0.0.1 port 46782 Nov 24 00:16:24.235990 sshd-session[5367]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:24.251333 systemd[1]: sshd@19-10.0.0.25:22-10.0.0.1:46782.service: Deactivated successfully. Nov 24 00:16:24.253746 systemd[1]: session-20.scope: Deactivated successfully. Nov 24 00:16:24.254606 systemd-logind[1519]: Session 20 logged out. Waiting for processes to exit. Nov 24 00:16:24.258080 systemd[1]: Started sshd@20-10.0.0.25:22-10.0.0.1:46790.service - OpenSSH per-connection server daemon (10.0.0.1:46790). Nov 24 00:16:24.258908 systemd-logind[1519]: Removed session 20. Nov 24 00:16:24.319173 sshd[5381]: Accepted publickey for core from 10.0.0.1 port 46790 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:24.321180 sshd-session[5381]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:24.328452 systemd-logind[1519]: New session 21 of user core. Nov 24 00:16:24.338833 systemd[1]: Started session-21.scope - Session 21 of User core. Nov 24 00:16:24.508828 sshd[5384]: Connection closed by 10.0.0.1 port 46790 Nov 24 00:16:24.509214 sshd-session[5381]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:24.514320 systemd[1]: sshd@20-10.0.0.25:22-10.0.0.1:46790.service: Deactivated successfully. Nov 24 00:16:24.516485 systemd[1]: session-21.scope: Deactivated successfully. Nov 24 00:16:24.517400 systemd-logind[1519]: Session 21 logged out. Waiting for processes to exit. Nov 24 00:16:24.519198 systemd-logind[1519]: Removed session 21. Nov 24 00:16:25.192201 kubelet[2683]: E1124 00:16:25.191721 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" podUID="024a87f6-fa17-474a-b3f9-39d35fa916e3" Nov 24 00:16:26.190588 kubelet[2683]: E1124 00:16:26.190526 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" podUID="3b61fcb5-22ca-4d05-817b-ade781b1a990" Nov 24 00:16:28.190782 kubelet[2683]: E1124 00:16:28.190735 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:29.191268 kubelet[2683]: E1124 00:16:29.191198 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-n2448" podUID="6cfea110-fbd7-44cd-904d-4c7d268ca542" Nov 24 00:16:29.528245 systemd[1]: Started sshd@21-10.0.0.25:22-10.0.0.1:46804.service - OpenSSH per-connection server daemon (10.0.0.1:46804). Nov 24 00:16:29.594312 sshd[5401]: Accepted publickey for core from 10.0.0.1 port 46804 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:29.596377 sshd-session[5401]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:29.602084 systemd-logind[1519]: New session 22 of user core. Nov 24 00:16:29.611963 systemd[1]: Started session-22.scope - Session 22 of User core. Nov 24 00:16:29.736537 sshd[5404]: Connection closed by 10.0.0.1 port 46804 Nov 24 00:16:29.736993 sshd-session[5401]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:29.742326 systemd[1]: sshd@21-10.0.0.25:22-10.0.0.1:46804.service: Deactivated successfully. Nov 24 00:16:29.744768 systemd[1]: session-22.scope: Deactivated successfully. Nov 24 00:16:29.745865 systemd-logind[1519]: Session 22 logged out. Waiting for processes to exit. Nov 24 00:16:29.747723 systemd-logind[1519]: Removed session 22. Nov 24 00:16:30.190821 kubelet[2683]: E1124 00:16:30.190759 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-hr2wl" podUID="22c38642-d71a-4cc4-996b-dcdaded5d62c" Nov 24 00:16:30.191691 kubelet[2683]: E1124 00:16:30.191643 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:16:31.191989 kubelet[2683]: E1124 00:16:31.191641 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 24 00:16:31.193215 containerd[1537]: time="2025-11-24T00:16:31.193166358Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 24 00:16:31.523393 containerd[1537]: time="2025-11-24T00:16:31.523220144Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:31.524731 containerd[1537]: time="2025-11-24T00:16:31.524645042Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 24 00:16:31.524806 containerd[1537]: time="2025-11-24T00:16:31.524654039Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 24 00:16:31.525131 kubelet[2683]: E1124 00:16:31.525033 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:16:31.525131 kubelet[2683]: E1124 00:16:31.525116 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 24 00:16:31.525310 kubelet[2683]: E1124 00:16:31.525260 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:ed3f3941752e478da85a0af9ae40cdb3,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4wwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-7657d947d7-4pbrs_calico-system(4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:31.527294 containerd[1537]: time="2025-11-24T00:16:31.527258120Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 24 00:16:31.878800 containerd[1537]: time="2025-11-24T00:16:31.878727537Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:31.924195 containerd[1537]: time="2025-11-24T00:16:31.924122174Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 24 00:16:31.924374 containerd[1537]: time="2025-11-24T00:16:31.924150078Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 24 00:16:31.924546 kubelet[2683]: E1124 00:16:31.924481 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:16:31.924677 kubelet[2683]: E1124 00:16:31.924554 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 24 00:16:31.924764 kubelet[2683]: E1124 00:16:31.924719 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-7657d947d7-4pbrs_calico-system(4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:31.925989 kubelet[2683]: E1124 00:16:31.925925 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-7657d947d7-4pbrs" podUID="4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2" Nov 24 00:16:34.751187 systemd[1]: Started sshd@22-10.0.0.25:22-10.0.0.1:43052.service - OpenSSH per-connection server daemon (10.0.0.1:43052). Nov 24 00:16:34.812042 sshd[5423]: Accepted publickey for core from 10.0.0.1 port 43052 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:34.814365 sshd-session[5423]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:34.819584 systemd-logind[1519]: New session 23 of user core. Nov 24 00:16:34.836013 systemd[1]: Started session-23.scope - Session 23 of User core. Nov 24 00:16:34.958291 sshd[5426]: Connection closed by 10.0.0.1 port 43052 Nov 24 00:16:34.958719 sshd-session[5423]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:34.963727 systemd[1]: sshd@22-10.0.0.25:22-10.0.0.1:43052.service: Deactivated successfully. Nov 24 00:16:34.965894 systemd[1]: session-23.scope: Deactivated successfully. Nov 24 00:16:34.966787 systemd-logind[1519]: Session 23 logged out. Waiting for processes to exit. Nov 24 00:16:34.968104 systemd-logind[1519]: Removed session 23. Nov 24 00:16:37.193528 containerd[1537]: time="2025-11-24T00:16:37.192803052Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 24 00:16:37.505658 containerd[1537]: time="2025-11-24T00:16:37.505454385Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:37.566262 containerd[1537]: time="2025-11-24T00:16:37.566161908Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 24 00:16:37.566460 containerd[1537]: time="2025-11-24T00:16:37.566240576Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 24 00:16:37.566647 kubelet[2683]: E1124 00:16:37.566563 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:16:37.567073 kubelet[2683]: E1124 00:16:37.566650 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 24 00:16:37.567073 kubelet[2683]: E1124 00:16:37.566831 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xn2qq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5889ff9675-8zv2t_calico-system(024a87f6-fa17-474a-b3f9-39d35fa916e3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:37.568436 kubelet[2683]: E1124 00:16:37.568372 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5889ff9675-8zv2t" podUID="024a87f6-fa17-474a-b3f9-39d35fa916e3" Nov 24 00:16:39.192809 containerd[1537]: time="2025-11-24T00:16:39.192424428Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:16:39.539352 containerd[1537]: time="2025-11-24T00:16:39.539284990Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:39.590053 containerd[1537]: time="2025-11-24T00:16:39.589960457Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:16:39.590053 containerd[1537]: time="2025-11-24T00:16:39.590022314Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:16:39.590420 kubelet[2683]: E1124 00:16:39.590324 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:39.590813 kubelet[2683]: E1124 00:16:39.590431 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:39.590813 kubelet[2683]: E1124 00:16:39.590560 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rbpm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-c779694df-9w4qh_calico-apiserver(3b61fcb5-22ca-4d05-817b-ade781b1a990): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:39.591791 kubelet[2683]: E1124 00:16:39.591726 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-9w4qh" podUID="3b61fcb5-22ca-4d05-817b-ade781b1a990" Nov 24 00:16:39.984023 systemd[1]: Started sshd@23-10.0.0.25:22-10.0.0.1:43060.service - OpenSSH per-connection server daemon (10.0.0.1:43060). Nov 24 00:16:40.037812 sshd[5441]: Accepted publickey for core from 10.0.0.1 port 43060 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:40.039699 sshd-session[5441]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:40.044750 systemd-logind[1519]: New session 24 of user core. Nov 24 00:16:40.060951 systemd[1]: Started session-24.scope - Session 24 of User core. Nov 24 00:16:40.177682 sshd[5444]: Connection closed by 10.0.0.1 port 43060 Nov 24 00:16:40.178055 sshd-session[5441]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:40.181701 systemd[1]: sshd@23-10.0.0.25:22-10.0.0.1:43060.service: Deactivated successfully. Nov 24 00:16:40.184201 systemd[1]: session-24.scope: Deactivated successfully. Nov 24 00:16:40.186944 systemd-logind[1519]: Session 24 logged out. Waiting for processes to exit. Nov 24 00:16:40.188258 systemd-logind[1519]: Removed session 24. Nov 24 00:16:40.191683 containerd[1537]: time="2025-11-24T00:16:40.191551358Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 24 00:16:40.529539 containerd[1537]: time="2025-11-24T00:16:40.529456391Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:40.530785 containerd[1537]: time="2025-11-24T00:16:40.530732813Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 24 00:16:40.530897 containerd[1537]: time="2025-11-24T00:16:40.530808758Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 24 00:16:40.531079 kubelet[2683]: E1124 00:16:40.531022 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:40.531139 kubelet[2683]: E1124 00:16:40.531090 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 24 00:16:40.531269 kubelet[2683]: E1124 00:16:40.531225 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hbz2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-c779694df-n2448_calico-apiserver(6cfea110-fbd7-44cd-904d-4c7d268ca542): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:40.532550 kubelet[2683]: E1124 00:16:40.532404 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-c779694df-n2448" podUID="6cfea110-fbd7-44cd-904d-4c7d268ca542" Nov 24 00:16:42.191082 containerd[1537]: time="2025-11-24T00:16:42.191025267Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 24 00:16:42.546021 containerd[1537]: time="2025-11-24T00:16:42.545956553Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:42.547293 containerd[1537]: time="2025-11-24T00:16:42.547254416Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 24 00:16:42.547394 containerd[1537]: time="2025-11-24T00:16:42.547347102Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 24 00:16:42.547538 kubelet[2683]: E1124 00:16:42.547488 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:16:42.548130 kubelet[2683]: E1124 00:16:42.547548 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 24 00:16:42.548180 containerd[1537]: time="2025-11-24T00:16:42.547892442Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 24 00:16:42.548220 kubelet[2683]: E1124 00:16:42.548155 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fkmrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:42.846077 containerd[1537]: time="2025-11-24T00:16:42.845906916Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:42.847442 containerd[1537]: time="2025-11-24T00:16:42.847377916Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 24 00:16:42.847442 containerd[1537]: time="2025-11-24T00:16:42.847409086Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 24 00:16:42.847727 kubelet[2683]: E1124 00:16:42.847667 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:16:42.847778 kubelet[2683]: E1124 00:16:42.847739 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 24 00:16:42.848571 kubelet[2683]: E1124 00:16:42.847993 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t59pv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-hr2wl_calico-system(22c38642-d71a-4cc4-996b-dcdaded5d62c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:42.848719 containerd[1537]: time="2025-11-24T00:16:42.848187787Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 24 00:16:42.849760 kubelet[2683]: E1124 00:16:42.849710 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-hr2wl" podUID="22c38642-d71a-4cc4-996b-dcdaded5d62c" Nov 24 00:16:43.183073 containerd[1537]: time="2025-11-24T00:16:43.182908426Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 24 00:16:43.219020 containerd[1537]: time="2025-11-24T00:16:43.218520371Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 24 00:16:43.219020 containerd[1537]: time="2025-11-24T00:16:43.218586546Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 24 00:16:43.219658 kubelet[2683]: E1124 00:16:43.219044 2683 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:16:43.219658 kubelet[2683]: E1124 00:16:43.219104 2683 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 24 00:16:43.219658 kubelet[2683]: E1124 00:16:43.219240 2683 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fkmrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-5jkmb_calico-system(e7cdfd74-3050-4e07-b3f8-0110308ddcf6): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 24 00:16:43.220944 kubelet[2683]: E1124 00:16:43.220866 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-5jkmb" podUID="e7cdfd74-3050-4e07-b3f8-0110308ddcf6" Nov 24 00:16:45.193455 systemd[1]: Started sshd@24-10.0.0.25:22-10.0.0.1:53806.service - OpenSSH per-connection server daemon (10.0.0.1:53806). Nov 24 00:16:45.248063 sshd[5457]: Accepted publickey for core from 10.0.0.1 port 53806 ssh2: RSA SHA256:BLlmoJVEAwNVcsQWPOPwU0WJtaKUh0hefjY8k+s4MOA Nov 24 00:16:45.250270 sshd-session[5457]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 24 00:16:45.257337 systemd-logind[1519]: New session 25 of user core. Nov 24 00:16:45.269977 systemd[1]: Started session-25.scope - Session 25 of User core. Nov 24 00:16:45.387163 sshd[5462]: Connection closed by 10.0.0.1 port 53806 Nov 24 00:16:45.387552 sshd-session[5457]: pam_unix(sshd:session): session closed for user core Nov 24 00:16:45.393042 systemd[1]: sshd@24-10.0.0.25:22-10.0.0.1:53806.service: Deactivated successfully. Nov 24 00:16:45.395429 systemd[1]: session-25.scope: Deactivated successfully. Nov 24 00:16:45.396318 systemd-logind[1519]: Session 25 logged out. Waiting for processes to exit. Nov 24 00:16:45.397611 systemd-logind[1519]: Removed session 25. Nov 24 00:16:46.197704 kubelet[2683]: E1124 00:16:46.197616 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-7657d947d7-4pbrs" podUID="4aa1d061-de13-4af8-bf9e-bd05bcbcd8d2"