Dec 12 18:38:24.831656 kernel: Linux version 6.12.61-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Fri Dec 12 15:21:28 -00 2025 Dec 12 18:38:24.831674 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=a214a2d85e162c493e8b13db2df50a43e1005a0e4854a1ae089a14f442a30022 Dec 12 18:38:24.831683 kernel: BIOS-provided physical RAM map: Dec 12 18:38:24.831688 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Dec 12 18:38:24.831693 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Dec 12 18:38:24.831697 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Dec 12 18:38:24.831703 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007cfdbfff] usable Dec 12 18:38:24.831708 kernel: BIOS-e820: [mem 0x000000007cfdc000-0x000000007cffffff] reserved Dec 12 18:38:24.831713 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Dec 12 18:38:24.831718 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Dec 12 18:38:24.831723 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Dec 12 18:38:24.831728 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Dec 12 18:38:24.831733 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Dec 12 18:38:24.831738 kernel: NX (Execute Disable) protection: active Dec 12 18:38:24.831744 kernel: APIC: Static calls initialized Dec 12 18:38:24.831750 kernel: SMBIOS 3.0.0 present. Dec 12 18:38:24.831755 kernel: DMI: Hetzner vServer/Standard PC (Q35 + ICH9, 2009), BIOS 20171111 11/11/2017 Dec 12 18:38:24.831761 kernel: DMI: Memory slots populated: 1/1 Dec 12 18:38:24.831766 kernel: Hypervisor detected: KVM Dec 12 18:38:24.831771 kernel: last_pfn = 0x7cfdc max_arch_pfn = 0x400000000 Dec 12 18:38:24.831776 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Dec 12 18:38:24.831781 kernel: kvm-clock: using sched offset of 4460664452 cycles Dec 12 18:38:24.831786 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Dec 12 18:38:24.831792 kernel: tsc: Detected 2445.406 MHz processor Dec 12 18:38:24.831797 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Dec 12 18:38:24.831804 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Dec 12 18:38:24.831810 kernel: last_pfn = 0x7cfdc max_arch_pfn = 0x400000000 Dec 12 18:38:24.831815 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Dec 12 18:38:24.831820 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Dec 12 18:38:24.831826 kernel: Using GB pages for direct mapping Dec 12 18:38:24.831831 kernel: ACPI: Early table checksum verification disabled Dec 12 18:38:24.831836 kernel: ACPI: RSDP 0x00000000000F5270 000014 (v00 BOCHS ) Dec 12 18:38:24.831841 kernel: ACPI: RSDT 0x000000007CFE2693 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:38:24.831847 kernel: ACPI: FACP 0x000000007CFE2483 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:38:24.831853 kernel: ACPI: DSDT 0x000000007CFE0040 002443 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:38:24.831859 kernel: ACPI: FACS 0x000000007CFE0000 000040 Dec 12 18:38:24.831864 kernel: ACPI: APIC 0x000000007CFE2577 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:38:24.831869 kernel: ACPI: HPET 0x000000007CFE25F7 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:38:24.831875 kernel: ACPI: MCFG 0x000000007CFE262F 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:38:24.831880 kernel: ACPI: WAET 0x000000007CFE266B 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:38:24.831888 kernel: ACPI: Reserving FACP table memory at [mem 0x7cfe2483-0x7cfe2576] Dec 12 18:38:24.831894 kernel: ACPI: Reserving DSDT table memory at [mem 0x7cfe0040-0x7cfe2482] Dec 12 18:38:24.831900 kernel: ACPI: Reserving FACS table memory at [mem 0x7cfe0000-0x7cfe003f] Dec 12 18:38:24.831906 kernel: ACPI: Reserving APIC table memory at [mem 0x7cfe2577-0x7cfe25f6] Dec 12 18:38:24.831911 kernel: ACPI: Reserving HPET table memory at [mem 0x7cfe25f7-0x7cfe262e] Dec 12 18:38:24.831917 kernel: ACPI: Reserving MCFG table memory at [mem 0x7cfe262f-0x7cfe266a] Dec 12 18:38:24.831922 kernel: ACPI: Reserving WAET table memory at [mem 0x7cfe266b-0x7cfe2692] Dec 12 18:38:24.831927 kernel: No NUMA configuration found Dec 12 18:38:24.831934 kernel: Faking a node at [mem 0x0000000000000000-0x000000007cfdbfff] Dec 12 18:38:24.831940 kernel: NODE_DATA(0) allocated [mem 0x7cfd4dc0-0x7cfdbfff] Dec 12 18:38:24.831945 kernel: Zone ranges: Dec 12 18:38:24.831951 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Dec 12 18:38:24.831956 kernel: DMA32 [mem 0x0000000001000000-0x000000007cfdbfff] Dec 12 18:38:24.831962 kernel: Normal empty Dec 12 18:38:24.831967 kernel: Device empty Dec 12 18:38:24.831972 kernel: Movable zone start for each node Dec 12 18:38:24.831978 kernel: Early memory node ranges Dec 12 18:38:24.831983 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Dec 12 18:38:24.831990 kernel: node 0: [mem 0x0000000000100000-0x000000007cfdbfff] Dec 12 18:38:24.831995 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007cfdbfff] Dec 12 18:38:24.832001 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 12 18:38:24.832006 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Dec 12 18:38:24.832012 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Dec 12 18:38:24.832017 kernel: ACPI: PM-Timer IO Port: 0x608 Dec 12 18:38:24.832023 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Dec 12 18:38:24.832028 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Dec 12 18:38:24.832034 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Dec 12 18:38:24.832040 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Dec 12 18:38:24.832046 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Dec 12 18:38:24.832051 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Dec 12 18:38:24.832057 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Dec 12 18:38:24.832062 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Dec 12 18:38:24.832068 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Dec 12 18:38:24.832073 kernel: CPU topo: Max. logical packages: 1 Dec 12 18:38:24.832079 kernel: CPU topo: Max. logical dies: 1 Dec 12 18:38:24.832084 kernel: CPU topo: Max. dies per package: 1 Dec 12 18:38:24.832090 kernel: CPU topo: Max. threads per core: 1 Dec 12 18:38:24.832096 kernel: CPU topo: Num. cores per package: 2 Dec 12 18:38:24.832102 kernel: CPU topo: Num. threads per package: 2 Dec 12 18:38:24.832107 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Dec 12 18:38:24.832112 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Dec 12 18:38:24.832118 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Dec 12 18:38:24.832124 kernel: Booting paravirtualized kernel on KVM Dec 12 18:38:24.832129 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Dec 12 18:38:24.832135 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Dec 12 18:38:24.832142 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Dec 12 18:38:24.832147 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Dec 12 18:38:24.832153 kernel: pcpu-alloc: [0] 0 1 Dec 12 18:38:24.832158 kernel: kvm-guest: PV spinlocks disabled, no host support Dec 12 18:38:24.832165 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=a214a2d85e162c493e8b13db2df50a43e1005a0e4854a1ae089a14f442a30022 Dec 12 18:38:24.832170 kernel: random: crng init done Dec 12 18:38:24.832176 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 12 18:38:24.832181 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Dec 12 18:38:24.832188 kernel: Fallback order for Node 0: 0 Dec 12 18:38:24.832194 kernel: Built 1 zonelists, mobility grouping on. Total pages: 511866 Dec 12 18:38:24.832199 kernel: Policy zone: DMA32 Dec 12 18:38:24.832205 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 12 18:38:24.832210 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Dec 12 18:38:24.832216 kernel: ftrace: allocating 40103 entries in 157 pages Dec 12 18:38:24.832221 kernel: ftrace: allocated 157 pages with 5 groups Dec 12 18:38:24.832227 kernel: Dynamic Preempt: voluntary Dec 12 18:38:24.832232 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 12 18:38:24.832238 kernel: rcu: RCU event tracing is enabled. Dec 12 18:38:24.832245 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Dec 12 18:38:24.832251 kernel: Trampoline variant of Tasks RCU enabled. Dec 12 18:38:24.832256 kernel: Rude variant of Tasks RCU enabled. Dec 12 18:38:24.832262 kernel: Tracing variant of Tasks RCU enabled. Dec 12 18:38:24.832267 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 12 18:38:24.832288 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Dec 12 18:38:24.832293 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 12 18:38:24.832299 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 12 18:38:24.832305 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 12 18:38:24.832312 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Dec 12 18:38:24.832317 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Dec 12 18:38:24.832323 kernel: Console: colour VGA+ 80x25 Dec 12 18:38:24.832328 kernel: printk: legacy console [tty0] enabled Dec 12 18:38:24.832348 kernel: printk: legacy console [ttyS0] enabled Dec 12 18:38:24.832355 kernel: ACPI: Core revision 20240827 Dec 12 18:38:24.832365 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Dec 12 18:38:24.832372 kernel: APIC: Switch to symmetric I/O mode setup Dec 12 18:38:24.832378 kernel: x2apic enabled Dec 12 18:38:24.832384 kernel: APIC: Switched APIC routing to: physical x2apic Dec 12 18:38:24.832390 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Dec 12 18:38:24.832396 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fc4eb620, max_idle_ns: 440795316590 ns Dec 12 18:38:24.832403 kernel: Calibrating delay loop (skipped) preset value.. 4890.81 BogoMIPS (lpj=2445406) Dec 12 18:38:24.832409 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Dec 12 18:38:24.832414 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Dec 12 18:38:24.832420 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Dec 12 18:38:24.832427 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Dec 12 18:38:24.832433 kernel: Spectre V2 : Mitigation: Retpolines Dec 12 18:38:24.832439 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Dec 12 18:38:24.832445 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Dec 12 18:38:24.832450 kernel: active return thunk: retbleed_return_thunk Dec 12 18:38:24.832456 kernel: RETBleed: Mitigation: untrained return thunk Dec 12 18:38:24.832462 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Dec 12 18:38:24.832468 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Dec 12 18:38:24.832474 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Dec 12 18:38:24.832481 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Dec 12 18:38:24.832486 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Dec 12 18:38:24.832492 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Dec 12 18:38:24.832498 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Dec 12 18:38:24.832504 kernel: Freeing SMP alternatives memory: 32K Dec 12 18:38:24.832510 kernel: pid_max: default: 32768 minimum: 301 Dec 12 18:38:24.832515 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Dec 12 18:38:24.832521 kernel: landlock: Up and running. Dec 12 18:38:24.832527 kernel: SELinux: Initializing. Dec 12 18:38:24.832534 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 12 18:38:24.832539 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 12 18:38:24.832545 kernel: smpboot: CPU0: AMD EPYC-Rome Processor (family: 0x17, model: 0x31, stepping: 0x0) Dec 12 18:38:24.832551 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Dec 12 18:38:24.832557 kernel: ... version: 0 Dec 12 18:38:24.832563 kernel: ... bit width: 48 Dec 12 18:38:24.832568 kernel: ... generic registers: 6 Dec 12 18:38:24.832574 kernel: ... value mask: 0000ffffffffffff Dec 12 18:38:24.832580 kernel: ... max period: 00007fffffffffff Dec 12 18:38:24.832587 kernel: ... fixed-purpose events: 0 Dec 12 18:38:24.832592 kernel: ... event mask: 000000000000003f Dec 12 18:38:24.832598 kernel: signal: max sigframe size: 1776 Dec 12 18:38:24.832604 kernel: rcu: Hierarchical SRCU implementation. Dec 12 18:38:24.832610 kernel: rcu: Max phase no-delay instances is 400. Dec 12 18:38:24.832616 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Dec 12 18:38:24.832621 kernel: smp: Bringing up secondary CPUs ... Dec 12 18:38:24.832627 kernel: smpboot: x86: Booting SMP configuration: Dec 12 18:38:24.832633 kernel: .... node #0, CPUs: #1 Dec 12 18:38:24.832640 kernel: smp: Brought up 1 node, 2 CPUs Dec 12 18:38:24.832645 kernel: smpboot: Total of 2 processors activated (9781.62 BogoMIPS) Dec 12 18:38:24.832652 kernel: Memory: 1909588K/2047464K available (14336K kernel code, 2444K rwdata, 26064K rodata, 46188K init, 2572K bss, 133332K reserved, 0K cma-reserved) Dec 12 18:38:24.832658 kernel: devtmpfs: initialized Dec 12 18:38:24.832664 kernel: x86/mm: Memory block size: 128MB Dec 12 18:38:24.832670 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 12 18:38:24.832676 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Dec 12 18:38:24.832682 kernel: pinctrl core: initialized pinctrl subsystem Dec 12 18:38:24.832688 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 12 18:38:24.832694 kernel: audit: initializing netlink subsys (disabled) Dec 12 18:38:24.832700 kernel: audit: type=2000 audit(1765564702.105:1): state=initialized audit_enabled=0 res=1 Dec 12 18:38:24.832706 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 12 18:38:24.832712 kernel: thermal_sys: Registered thermal governor 'user_space' Dec 12 18:38:24.832717 kernel: cpuidle: using governor menu Dec 12 18:38:24.832723 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 12 18:38:24.832729 kernel: dca service started, version 1.12.1 Dec 12 18:38:24.832735 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] (base 0xb0000000) for domain 0000 [bus 00-ff] Dec 12 18:38:24.832740 kernel: PCI: Using configuration type 1 for base access Dec 12 18:38:24.832747 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Dec 12 18:38:24.832753 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Dec 12 18:38:24.832759 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Dec 12 18:38:24.832764 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Dec 12 18:38:24.832770 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Dec 12 18:38:24.832776 kernel: ACPI: Added _OSI(Module Device) Dec 12 18:38:24.832782 kernel: ACPI: Added _OSI(Processor Device) Dec 12 18:38:24.832787 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 12 18:38:24.832793 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 12 18:38:24.832804 kernel: ACPI: Interpreter enabled Dec 12 18:38:24.832814 kernel: ACPI: PM: (supports S0 S5) Dec 12 18:38:24.832825 kernel: ACPI: Using IOAPIC for interrupt routing Dec 12 18:38:24.832836 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Dec 12 18:38:24.832846 kernel: PCI: Using E820 reservations for host bridge windows Dec 12 18:38:24.832857 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Dec 12 18:38:24.832868 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 12 18:38:24.832982 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Dec 12 18:38:24.833074 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Dec 12 18:38:24.833135 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Dec 12 18:38:24.833143 kernel: PCI host bridge to bus 0000:00 Dec 12 18:38:24.833206 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Dec 12 18:38:24.833259 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Dec 12 18:38:24.833403 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Dec 12 18:38:24.833463 kernel: pci_bus 0000:00: root bus resource [mem 0x7d000000-0xafffffff window] Dec 12 18:38:24.833518 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Dec 12 18:38:24.833566 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Dec 12 18:38:24.833614 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 12 18:38:24.833685 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Dec 12 18:38:24.833755 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x030000 conventional PCI endpoint Dec 12 18:38:24.833816 kernel: pci 0000:00:01.0: BAR 0 [mem 0xfb800000-0xfbffffff pref] Dec 12 18:38:24.833877 kernel: pci 0000:00:01.0: BAR 2 [mem 0xfd200000-0xfd203fff 64bit pref] Dec 12 18:38:24.833934 kernel: pci 0000:00:01.0: BAR 4 [mem 0xfea10000-0xfea10fff] Dec 12 18:38:24.833989 kernel: pci 0000:00:01.0: ROM [mem 0xfea00000-0xfea0ffff pref] Dec 12 18:38:24.834044 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Dec 12 18:38:24.834107 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.834163 kernel: pci 0000:00:02.0: BAR 0 [mem 0xfea11000-0xfea11fff] Dec 12 18:38:24.834219 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 12 18:38:24.834292 kernel: pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff] Dec 12 18:38:24.836388 kernel: pci 0000:00:02.0: bridge window [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 12 18:38:24.836477 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.836542 kernel: pci 0000:00:02.1: BAR 0 [mem 0xfea12000-0xfea12fff] Dec 12 18:38:24.836600 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 12 18:38:24.836696 kernel: pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff] Dec 12 18:38:24.836767 kernel: pci 0000:00:02.1: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Dec 12 18:38:24.836837 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.836896 kernel: pci 0000:00:02.2: BAR 0 [mem 0xfea13000-0xfea13fff] Dec 12 18:38:24.836958 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 12 18:38:24.837068 kernel: pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff] Dec 12 18:38:24.837136 kernel: pci 0000:00:02.2: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 12 18:38:24.837204 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.837264 kernel: pci 0000:00:02.3: BAR 0 [mem 0xfea14000-0xfea14fff] Dec 12 18:38:24.837367 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 12 18:38:24.837430 kernel: pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff] Dec 12 18:38:24.837486 kernel: pci 0000:00:02.3: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 12 18:38:24.837549 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.837613 kernel: pci 0000:00:02.4: BAR 0 [mem 0xfea15000-0xfea15fff] Dec 12 18:38:24.837714 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 12 18:38:24.837801 kernel: pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff] Dec 12 18:38:24.837864 kernel: pci 0000:00:02.4: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 12 18:38:24.837931 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.837989 kernel: pci 0000:00:02.5: BAR 0 [mem 0xfea16000-0xfea16fff] Dec 12 18:38:24.838067 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 12 18:38:24.838173 kernel: pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff] Dec 12 18:38:24.838305 kernel: pci 0000:00:02.5: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 12 18:38:24.840451 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.840529 kernel: pci 0000:00:02.6: BAR 0 [mem 0xfea17000-0xfea17fff] Dec 12 18:38:24.840590 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 12 18:38:24.840648 kernel: pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff] Dec 12 18:38:24.840704 kernel: pci 0000:00:02.6: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 12 18:38:24.840767 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.840825 kernel: pci 0000:00:02.7: BAR 0 [mem 0xfea18000-0xfea18fff] Dec 12 18:38:24.840886 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 12 18:38:24.840942 kernel: pci 0000:00:02.7: bridge window [mem 0xfda00000-0xfdbfffff] Dec 12 18:38:24.840998 kernel: pci 0000:00:02.7: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 12 18:38:24.841059 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 12 18:38:24.841139 kernel: pci 0000:00:03.0: BAR 0 [mem 0xfea19000-0xfea19fff] Dec 12 18:38:24.841244 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 12 18:38:24.841325 kernel: pci 0000:00:03.0: bridge window [mem 0xfd800000-0xfd9fffff] Dec 12 18:38:24.841438 kernel: pci 0000:00:03.0: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 12 18:38:24.841511 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Dec 12 18:38:24.841570 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Dec 12 18:38:24.841634 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Dec 12 18:38:24.841690 kernel: pci 0000:00:1f.2: BAR 4 [io 0xc040-0xc05f] Dec 12 18:38:24.841745 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xfea1a000-0xfea1afff] Dec 12 18:38:24.841805 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Dec 12 18:38:24.841865 kernel: pci 0000:00:1f.3: BAR 4 [io 0x0700-0x073f] Dec 12 18:38:24.841931 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 PCIe Endpoint Dec 12 18:38:24.841990 kernel: pci 0000:01:00.0: BAR 1 [mem 0xfe880000-0xfe880fff] Dec 12 18:38:24.842059 kernel: pci 0000:01:00.0: BAR 4 [mem 0xfd000000-0xfd003fff 64bit pref] Dec 12 18:38:24.842158 kernel: pci 0000:01:00.0: ROM [mem 0xfe800000-0xfe87ffff pref] Dec 12 18:38:24.842251 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 12 18:38:24.842365 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 PCIe Endpoint Dec 12 18:38:24.842442 kernel: pci 0000:02:00.0: BAR 0 [mem 0xfe600000-0xfe603fff 64bit] Dec 12 18:38:24.842505 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 12 18:38:24.842577 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 PCIe Endpoint Dec 12 18:38:24.842641 kernel: pci 0000:03:00.0: BAR 1 [mem 0xfe400000-0xfe400fff] Dec 12 18:38:24.842703 kernel: pci 0000:03:00.0: BAR 4 [mem 0xfcc00000-0xfcc03fff 64bit pref] Dec 12 18:38:24.842763 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 12 18:38:24.842829 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 PCIe Endpoint Dec 12 18:38:24.842918 kernel: pci 0000:04:00.0: BAR 4 [mem 0xfca00000-0xfca03fff 64bit pref] Dec 12 18:38:24.843017 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 12 18:38:24.843091 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 PCIe Endpoint Dec 12 18:38:24.843157 kernel: pci 0000:05:00.0: BAR 1 [mem 0xfe000000-0xfe000fff] Dec 12 18:38:24.843217 kernel: pci 0000:05:00.0: BAR 4 [mem 0xfc800000-0xfc803fff 64bit pref] Dec 12 18:38:24.843289 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 12 18:38:24.843438 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 PCIe Endpoint Dec 12 18:38:24.843520 kernel: pci 0000:06:00.0: BAR 1 [mem 0xfde00000-0xfde00fff] Dec 12 18:38:24.843586 kernel: pci 0000:06:00.0: BAR 4 [mem 0xfc600000-0xfc603fff 64bit pref] Dec 12 18:38:24.843647 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 12 18:38:24.843656 kernel: acpiphp: Slot [0] registered Dec 12 18:38:24.843723 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 PCIe Endpoint Dec 12 18:38:24.843788 kernel: pci 0000:07:00.0: BAR 1 [mem 0xfdc80000-0xfdc80fff] Dec 12 18:38:24.843855 kernel: pci 0000:07:00.0: BAR 4 [mem 0xfc400000-0xfc403fff 64bit pref] Dec 12 18:38:24.843918 kernel: pci 0000:07:00.0: ROM [mem 0xfdc00000-0xfdc7ffff pref] Dec 12 18:38:24.843976 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 12 18:38:24.843986 kernel: acpiphp: Slot [0-2] registered Dec 12 18:38:24.844043 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 12 18:38:24.844052 kernel: acpiphp: Slot [0-3] registered Dec 12 18:38:24.844107 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 12 18:38:24.844116 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Dec 12 18:38:24.844125 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Dec 12 18:38:24.844131 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Dec 12 18:38:24.844137 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Dec 12 18:38:24.844143 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Dec 12 18:38:24.844149 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Dec 12 18:38:24.844155 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Dec 12 18:38:24.844161 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Dec 12 18:38:24.844166 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Dec 12 18:38:24.844172 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Dec 12 18:38:24.844179 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Dec 12 18:38:24.844185 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Dec 12 18:38:24.844191 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Dec 12 18:38:24.844196 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Dec 12 18:38:24.844202 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Dec 12 18:38:24.844208 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Dec 12 18:38:24.844214 kernel: iommu: Default domain type: Translated Dec 12 18:38:24.844220 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Dec 12 18:38:24.844226 kernel: PCI: Using ACPI for IRQ routing Dec 12 18:38:24.844233 kernel: PCI: pci_cache_line_size set to 64 bytes Dec 12 18:38:24.844239 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Dec 12 18:38:24.844245 kernel: e820: reserve RAM buffer [mem 0x7cfdc000-0x7fffffff] Dec 12 18:38:24.844324 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Dec 12 18:38:24.844417 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Dec 12 18:38:24.844479 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Dec 12 18:38:24.844488 kernel: vgaarb: loaded Dec 12 18:38:24.844494 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Dec 12 18:38:24.844503 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Dec 12 18:38:24.844509 kernel: clocksource: Switched to clocksource kvm-clock Dec 12 18:38:24.844515 kernel: VFS: Disk quotas dquot_6.6.0 Dec 12 18:38:24.844521 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 12 18:38:24.844527 kernel: pnp: PnP ACPI init Dec 12 18:38:24.844594 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Dec 12 18:38:24.844605 kernel: pnp: PnP ACPI: found 5 devices Dec 12 18:38:24.844611 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Dec 12 18:38:24.844617 kernel: NET: Registered PF_INET protocol family Dec 12 18:38:24.844625 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 12 18:38:24.844631 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Dec 12 18:38:24.844637 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 12 18:38:24.844643 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Dec 12 18:38:24.844649 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Dec 12 18:38:24.844655 kernel: TCP: Hash tables configured (established 16384 bind 16384) Dec 12 18:38:24.844661 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 12 18:38:24.844667 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 12 18:38:24.844674 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 12 18:38:24.844680 kernel: NET: Registered PF_XDP protocol family Dec 12 18:38:24.844741 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Dec 12 18:38:24.844803 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Dec 12 18:38:24.844865 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Dec 12 18:38:24.844924 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff]: assigned Dec 12 18:38:24.844985 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff]: assigned Dec 12 18:38:24.845111 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff]: assigned Dec 12 18:38:24.845185 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 12 18:38:24.845247 kernel: pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff] Dec 12 18:38:24.845329 kernel: pci 0000:00:02.0: bridge window [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 12 18:38:24.845417 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 12 18:38:24.845478 kernel: pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff] Dec 12 18:38:24.845566 kernel: pci 0000:00:02.1: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Dec 12 18:38:24.845641 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 12 18:38:24.845703 kernel: pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff] Dec 12 18:38:24.845764 kernel: pci 0000:00:02.2: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 12 18:38:24.845823 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 12 18:38:24.845889 kernel: pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff] Dec 12 18:38:24.845952 kernel: pci 0000:00:02.3: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 12 18:38:24.846010 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 12 18:38:24.846068 kernel: pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff] Dec 12 18:38:24.846125 kernel: pci 0000:00:02.4: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 12 18:38:24.846201 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 12 18:38:24.846297 kernel: pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff] Dec 12 18:38:24.846400 kernel: pci 0000:00:02.5: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 12 18:38:24.846465 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 12 18:38:24.846526 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff] Dec 12 18:38:24.846588 kernel: pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff] Dec 12 18:38:24.846649 kernel: pci 0000:00:02.6: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 12 18:38:24.846713 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 12 18:38:24.846811 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff] Dec 12 18:38:24.846888 kernel: pci 0000:00:02.7: bridge window [mem 0xfda00000-0xfdbfffff] Dec 12 18:38:24.846948 kernel: pci 0000:00:02.7: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 12 18:38:24.847009 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 12 18:38:24.847067 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff] Dec 12 18:38:24.847153 kernel: pci 0000:00:03.0: bridge window [mem 0xfd800000-0xfd9fffff] Dec 12 18:38:24.847231 kernel: pci 0000:00:03.0: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 12 18:38:24.847353 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Dec 12 18:38:24.847423 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Dec 12 18:38:24.847477 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Dec 12 18:38:24.847528 kernel: pci_bus 0000:00: resource 7 [mem 0x7d000000-0xafffffff window] Dec 12 18:38:24.847581 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Dec 12 18:38:24.847639 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Dec 12 18:38:24.847702 kernel: pci_bus 0000:01: resource 1 [mem 0xfe800000-0xfe9fffff] Dec 12 18:38:24.847759 kernel: pci_bus 0000:01: resource 2 [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 12 18:38:24.847824 kernel: pci_bus 0000:02: resource 1 [mem 0xfe600000-0xfe7fffff] Dec 12 18:38:24.847881 kernel: pci_bus 0000:02: resource 2 [mem 0xfce00000-0xfcffffff 64bit pref] Dec 12 18:38:24.847945 kernel: pci_bus 0000:03: resource 1 [mem 0xfe400000-0xfe5fffff] Dec 12 18:38:24.848003 kernel: pci_bus 0000:03: resource 2 [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 12 18:38:24.848061 kernel: pci_bus 0000:04: resource 1 [mem 0xfe200000-0xfe3fffff] Dec 12 18:38:24.848116 kernel: pci_bus 0000:04: resource 2 [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 12 18:38:24.848181 kernel: pci_bus 0000:05: resource 1 [mem 0xfe000000-0xfe1fffff] Dec 12 18:38:24.848237 kernel: pci_bus 0000:05: resource 2 [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 12 18:38:24.848325 kernel: pci_bus 0000:06: resource 1 [mem 0xfde00000-0xfdffffff] Dec 12 18:38:24.848407 kernel: pci_bus 0000:06: resource 2 [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 12 18:38:24.848474 kernel: pci_bus 0000:07: resource 0 [io 0x1000-0x1fff] Dec 12 18:38:24.848531 kernel: pci_bus 0000:07: resource 1 [mem 0xfdc00000-0xfddfffff] Dec 12 18:38:24.848585 kernel: pci_bus 0000:07: resource 2 [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 12 18:38:24.848650 kernel: pci_bus 0000:08: resource 0 [io 0x2000-0x2fff] Dec 12 18:38:24.848706 kernel: pci_bus 0000:08: resource 1 [mem 0xfda00000-0xfdbfffff] Dec 12 18:38:24.848759 kernel: pci_bus 0000:08: resource 2 [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 12 18:38:24.848818 kernel: pci_bus 0000:09: resource 0 [io 0x3000-0x3fff] Dec 12 18:38:24.848874 kernel: pci_bus 0000:09: resource 1 [mem 0xfd800000-0xfd9fffff] Dec 12 18:38:24.848926 kernel: pci_bus 0000:09: resource 2 [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 12 18:38:24.848937 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Dec 12 18:38:24.848944 kernel: PCI: CLS 0 bytes, default 64 Dec 12 18:38:24.848951 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fc4eb620, max_idle_ns: 440795316590 ns Dec 12 18:38:24.848957 kernel: Initialise system trusted keyrings Dec 12 18:38:24.848963 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Dec 12 18:38:24.848970 kernel: Key type asymmetric registered Dec 12 18:38:24.848976 kernel: Asymmetric key parser 'x509' registered Dec 12 18:38:24.848982 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Dec 12 18:38:24.848988 kernel: io scheduler mq-deadline registered Dec 12 18:38:24.848995 kernel: io scheduler kyber registered Dec 12 18:38:24.849001 kernel: io scheduler bfq registered Dec 12 18:38:24.849062 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Dec 12 18:38:24.849122 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Dec 12 18:38:24.849190 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Dec 12 18:38:24.849290 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Dec 12 18:38:24.849729 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Dec 12 18:38:24.849804 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Dec 12 18:38:24.849866 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Dec 12 18:38:24.849932 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Dec 12 18:38:24.849992 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Dec 12 18:38:24.850091 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Dec 12 18:38:24.850161 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Dec 12 18:38:24.850220 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Dec 12 18:38:24.850294 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Dec 12 18:38:24.850377 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Dec 12 18:38:24.850445 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Dec 12 18:38:24.850550 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Dec 12 18:38:24.850569 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Dec 12 18:38:24.850643 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 32 Dec 12 18:38:24.850706 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 32 Dec 12 18:38:24.850716 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Dec 12 18:38:24.850723 kernel: ACPI: \_SB_.GSIF: Enabled at IRQ 21 Dec 12 18:38:24.850733 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 12 18:38:24.850739 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Dec 12 18:38:24.850745 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Dec 12 18:38:24.850752 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Dec 12 18:38:24.850758 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Dec 12 18:38:24.850764 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Dec 12 18:38:24.850829 kernel: rtc_cmos 00:03: RTC can wake from S4 Dec 12 18:38:24.850886 kernel: rtc_cmos 00:03: registered as rtc0 Dec 12 18:38:24.850943 kernel: rtc_cmos 00:03: setting system clock to 2025-12-12T18:38:24 UTC (1765564704) Dec 12 18:38:24.850996 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Dec 12 18:38:24.851005 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Dec 12 18:38:24.851011 kernel: NET: Registered PF_INET6 protocol family Dec 12 18:38:24.851017 kernel: Segment Routing with IPv6 Dec 12 18:38:24.851024 kernel: In-situ OAM (IOAM) with IPv6 Dec 12 18:38:24.851030 kernel: NET: Registered PF_PACKET protocol family Dec 12 18:38:24.851036 kernel: Key type dns_resolver registered Dec 12 18:38:24.851045 kernel: IPI shorthand broadcast: enabled Dec 12 18:38:24.851051 kernel: sched_clock: Marking stable (3232007437, 263313279)->(3527222261, -31901545) Dec 12 18:38:24.851057 kernel: registered taskstats version 1 Dec 12 18:38:24.851063 kernel: Loading compiled-in X.509 certificates Dec 12 18:38:24.851070 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.61-flatcar: 0d0c78e6590cb40d27f1cef749ef9f2f3425f38d' Dec 12 18:38:24.851076 kernel: Demotion targets for Node 0: null Dec 12 18:38:24.851082 kernel: Key type .fscrypt registered Dec 12 18:38:24.851088 kernel: Key type fscrypt-provisioning registered Dec 12 18:38:24.851094 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 12 18:38:24.851101 kernel: ima: Allocated hash algorithm: sha1 Dec 12 18:38:24.851107 kernel: ima: No architecture policies found Dec 12 18:38:24.851113 kernel: clk: Disabling unused clocks Dec 12 18:38:24.851119 kernel: Warning: unable to open an initial console. Dec 12 18:38:24.851126 kernel: Freeing unused kernel image (initmem) memory: 46188K Dec 12 18:38:24.851132 kernel: Write protecting the kernel read-only data: 40960k Dec 12 18:38:24.851138 kernel: Freeing unused kernel image (rodata/data gap) memory: 560K Dec 12 18:38:24.851144 kernel: Run /init as init process Dec 12 18:38:24.851150 kernel: with arguments: Dec 12 18:38:24.851158 kernel: /init Dec 12 18:38:24.851164 kernel: with environment: Dec 12 18:38:24.851170 kernel: HOME=/ Dec 12 18:38:24.851176 kernel: TERM=linux Dec 12 18:38:24.851184 systemd[1]: Successfully made /usr/ read-only. Dec 12 18:38:24.851193 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Dec 12 18:38:24.851200 systemd[1]: Detected virtualization kvm. Dec 12 18:38:24.851207 systemd[1]: Detected architecture x86-64. Dec 12 18:38:24.851214 systemd[1]: Running in initrd. Dec 12 18:38:24.851221 systemd[1]: No hostname configured, using default hostname. Dec 12 18:38:24.851227 systemd[1]: Hostname set to . Dec 12 18:38:24.851234 systemd[1]: Initializing machine ID from VM UUID. Dec 12 18:38:24.851240 systemd[1]: Queued start job for default target initrd.target. Dec 12 18:38:24.851247 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 12 18:38:24.851253 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 12 18:38:24.851260 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Dec 12 18:38:24.851268 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 12 18:38:24.851291 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Dec 12 18:38:24.851298 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Dec 12 18:38:24.851306 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Dec 12 18:38:24.851313 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Dec 12 18:38:24.851319 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 12 18:38:24.851326 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 12 18:38:24.851347 systemd[1]: Reached target paths.target - Path Units. Dec 12 18:38:24.851354 systemd[1]: Reached target slices.target - Slice Units. Dec 12 18:38:24.851360 systemd[1]: Reached target swap.target - Swaps. Dec 12 18:38:24.851367 systemd[1]: Reached target timers.target - Timer Units. Dec 12 18:38:24.851373 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Dec 12 18:38:24.851380 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 12 18:38:24.851387 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 12 18:38:24.851393 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Dec 12 18:38:24.851400 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 12 18:38:24.851408 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 12 18:38:24.851415 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 12 18:38:24.851422 systemd[1]: Reached target sockets.target - Socket Units. Dec 12 18:38:24.851428 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Dec 12 18:38:24.851435 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 12 18:38:24.851441 systemd[1]: Finished network-cleanup.service - Network Cleanup. Dec 12 18:38:24.851448 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Dec 12 18:38:24.851455 systemd[1]: Starting systemd-fsck-usr.service... Dec 12 18:38:24.851462 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 12 18:38:24.851469 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 12 18:38:24.851492 systemd-journald[198]: Collecting audit messages is disabled. Dec 12 18:38:24.851510 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:38:24.851519 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Dec 12 18:38:24.851527 systemd-journald[198]: Journal started Dec 12 18:38:24.851544 systemd-journald[198]: Runtime Journal (/run/log/journal/6bc877b412ba488ab1eb8306b72c236b) is 4.7M, max 38.3M, 33.5M free. Dec 12 18:38:24.853515 systemd-modules-load[199]: Inserted module 'overlay' Dec 12 18:38:24.857007 systemd[1]: Started systemd-journald.service - Journal Service. Dec 12 18:38:24.859046 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 12 18:38:24.860601 systemd[1]: Finished systemd-fsck-usr.service. Dec 12 18:38:24.863085 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 12 18:38:24.866427 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 12 18:38:24.882485 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 12 18:38:24.957423 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 12 18:38:24.957451 kernel: Bridge firewalling registered Dec 12 18:38:24.886794 systemd-tmpfiles[210]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Dec 12 18:38:24.888173 systemd-modules-load[199]: Inserted module 'br_netfilter' Dec 12 18:38:24.958304 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 12 18:38:24.959801 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:38:24.961296 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 12 18:38:24.963970 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 12 18:38:24.976834 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 12 18:38:24.979000 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 12 18:38:24.986626 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 12 18:38:24.988631 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 12 18:38:24.992201 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 12 18:38:24.994571 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 12 18:38:24.996688 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Dec 12 18:38:25.017890 dracut-cmdline[239]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=a214a2d85e162c493e8b13db2df50a43e1005a0e4854a1ae089a14f442a30022 Dec 12 18:38:25.020536 systemd-resolved[236]: Positive Trust Anchors: Dec 12 18:38:25.020543 systemd-resolved[236]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 12 18:38:25.020567 systemd-resolved[236]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 12 18:38:25.022722 systemd-resolved[236]: Defaulting to hostname 'linux'. Dec 12 18:38:25.023559 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 12 18:38:25.024567 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 12 18:38:25.083421 kernel: SCSI subsystem initialized Dec 12 18:38:25.092387 kernel: Loading iSCSI transport class v2.0-870. Dec 12 18:38:25.103363 kernel: iscsi: registered transport (tcp) Dec 12 18:38:25.125147 kernel: iscsi: registered transport (qla4xxx) Dec 12 18:38:25.125195 kernel: QLogic iSCSI HBA Driver Dec 12 18:38:25.139472 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 12 18:38:25.150289 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 12 18:38:25.153020 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 12 18:38:25.182058 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Dec 12 18:38:25.183959 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Dec 12 18:38:25.229382 kernel: raid6: avx2x4 gen() 31548 MB/s Dec 12 18:38:25.247379 kernel: raid6: avx2x2 gen() 35187 MB/s Dec 12 18:38:25.266466 kernel: raid6: avx2x1 gen() 24864 MB/s Dec 12 18:38:25.266510 kernel: raid6: using algorithm avx2x2 gen() 35187 MB/s Dec 12 18:38:25.286481 kernel: raid6: .... xor() 32039 MB/s, rmw enabled Dec 12 18:38:25.286523 kernel: raid6: using avx2x2 recovery algorithm Dec 12 18:38:25.304371 kernel: xor: automatically using best checksumming function avx Dec 12 18:38:25.431387 kernel: Btrfs loaded, zoned=no, fsverity=no Dec 12 18:38:25.436856 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Dec 12 18:38:25.438743 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 12 18:38:25.466376 systemd-udevd[447]: Using default interface naming scheme 'v255'. Dec 12 18:38:25.471966 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 12 18:38:25.476442 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Dec 12 18:38:25.496712 dracut-pre-trigger[455]: rd.md=0: removing MD RAID activation Dec 12 18:38:25.517004 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Dec 12 18:38:25.519325 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 12 18:38:25.572956 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 12 18:38:25.575468 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Dec 12 18:38:25.654361 kernel: cryptd: max_cpu_qlen set to 1000 Dec 12 18:38:25.659386 kernel: virtio_scsi virtio5: 2/0/0 default/read/poll queues Dec 12 18:38:25.679440 kernel: scsi host0: Virtio SCSI HBA Dec 12 18:38:25.685929 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Dec 12 18:38:25.698608 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Dec 12 18:38:25.698008 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 12 18:38:25.698165 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:38:25.725449 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:38:25.727094 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:38:25.734358 kernel: ACPI: bus type USB registered Dec 12 18:38:25.742633 kernel: usbcore: registered new interface driver usbfs Dec 12 18:38:25.742760 kernel: AES CTR mode by8 optimization enabled Dec 12 18:38:25.745536 kernel: usbcore: registered new interface driver hub Dec 12 18:38:25.749359 kernel: usbcore: registered new device driver usb Dec 12 18:38:25.751353 kernel: libata version 3.00 loaded. Dec 12 18:38:25.772984 kernel: sd 0:0:0:0: Power-on or device reset occurred Dec 12 18:38:25.773204 kernel: sd 0:0:0:0: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Dec 12 18:38:25.773332 kernel: sd 0:0:0:0: [sda] Write Protect is off Dec 12 18:38:25.773475 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Dec 12 18:38:25.773564 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Dec 12 18:38:25.773650 kernel: ahci 0000:00:1f.2: version 3.0 Dec 12 18:38:25.773745 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Dec 12 18:38:25.773756 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Dec 12 18:38:25.773837 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Dec 12 18:38:25.773913 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Dec 12 18:38:25.777360 kernel: scsi host1: ahci Dec 12 18:38:25.777496 kernel: scsi host2: ahci Dec 12 18:38:25.778431 kernel: scsi host3: ahci Dec 12 18:38:25.778560 kernel: scsi host4: ahci Dec 12 18:38:25.779389 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 12 18:38:25.779462 kernel: GPT:17805311 != 80003071 Dec 12 18:38:25.779497 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 12 18:38:25.779525 kernel: GPT:17805311 != 80003071 Dec 12 18:38:25.779553 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 12 18:38:25.779581 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 12 18:38:25.779625 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Dec 12 18:38:25.779988 kernel: scsi host5: ahci Dec 12 18:38:25.782628 kernel: scsi host6: ahci Dec 12 18:38:25.782732 kernel: ata1: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a100 irq 48 lpm-pol 1 Dec 12 18:38:25.782742 kernel: ata2: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a180 irq 48 lpm-pol 1 Dec 12 18:38:25.782750 kernel: ata3: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a200 irq 48 lpm-pol 1 Dec 12 18:38:25.782758 kernel: ata4: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a280 irq 48 lpm-pol 1 Dec 12 18:38:25.782769 kernel: ata5: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a300 irq 48 lpm-pol 1 Dec 12 18:38:25.782776 kernel: ata6: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a380 irq 48 lpm-pol 1 Dec 12 18:38:25.842683 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Dec 12 18:38:25.922179 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:38:25.933058 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Dec 12 18:38:25.943973 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Dec 12 18:38:25.952499 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Dec 12 18:38:25.953371 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Dec 12 18:38:25.962228 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Dec 12 18:38:25.988023 disk-uuid[606]: Primary Header is updated. Dec 12 18:38:25.988023 disk-uuid[606]: Secondary Entries is updated. Dec 12 18:38:25.988023 disk-uuid[606]: Secondary Header is updated. Dec 12 18:38:26.003378 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 12 18:38:26.091880 kernel: ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Dec 12 18:38:26.099726 kernel: ata1.00: LPM support broken, forcing max_power Dec 12 18:38:26.099763 kernel: ata1.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Dec 12 18:38:26.099773 kernel: ata1.00: applying bridge limits Dec 12 18:38:26.099781 kernel: ata5: SATA link down (SStatus 0 SControl 300) Dec 12 18:38:26.103663 kernel: ata3: SATA link down (SStatus 0 SControl 300) Dec 12 18:38:26.103697 kernel: ata4: SATA link down (SStatus 0 SControl 300) Dec 12 18:38:26.105348 kernel: ata6: SATA link down (SStatus 0 SControl 300) Dec 12 18:38:26.107582 kernel: ata2: SATA link down (SStatus 0 SControl 300) Dec 12 18:38:26.112707 kernel: ata1.00: LPM support broken, forcing max_power Dec 12 18:38:26.112730 kernel: ata1.00: configured for UDMA/100 Dec 12 18:38:26.115706 kernel: scsi 1:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Dec 12 18:38:26.138814 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Dec 12 18:38:26.142368 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Dec 12 18:38:26.142498 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Dec 12 18:38:26.150084 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Dec 12 18:38:26.150196 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Dec 12 18:38:26.150291 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Dec 12 18:38:26.152607 kernel: hub 1-0:1.0: USB hub found Dec 12 18:38:26.154904 kernel: hub 1-0:1.0: 4 ports detected Dec 12 18:38:26.155819 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Dec 12 18:38:26.159127 kernel: hub 2-0:1.0: USB hub found Dec 12 18:38:26.159240 kernel: hub 2-0:1.0: 4 ports detected Dec 12 18:38:26.164555 kernel: sr 1:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Dec 12 18:38:26.164670 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Dec 12 18:38:26.180401 kernel: sr 1:0:0:0: Attached scsi CD-ROM sr0 Dec 12 18:38:26.395563 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Dec 12 18:38:26.482598 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Dec 12 18:38:26.483831 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Dec 12 18:38:26.485096 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 12 18:38:26.486822 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 12 18:38:26.489221 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Dec 12 18:38:26.515499 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Dec 12 18:38:26.531535 kernel: hid: raw HID events driver (C) Jiri Kosina Dec 12 18:38:26.538070 kernel: usbcore: registered new interface driver usbhid Dec 12 18:38:26.538103 kernel: usbhid: USB HID core driver Dec 12 18:38:26.545688 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input3 Dec 12 18:38:26.545726 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Dec 12 18:38:27.018974 disk-uuid[607]: The operation has completed successfully. Dec 12 18:38:27.021419 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 12 18:38:27.061307 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 12 18:38:27.061419 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Dec 12 18:38:27.095371 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Dec 12 18:38:27.112295 sh[641]: Success Dec 12 18:38:27.132551 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 12 18:38:27.132596 kernel: device-mapper: uevent: version 1.0.3 Dec 12 18:38:27.133387 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Dec 12 18:38:27.145384 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Dec 12 18:38:27.188803 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Dec 12 18:38:27.192407 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Dec 12 18:38:27.203780 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Dec 12 18:38:27.217365 kernel: BTRFS: device fsid a6ae7f96-a076-4d3c-81ed-46dd341492f8 devid 1 transid 37 /dev/mapper/usr (254:0) scanned by mount (653) Dec 12 18:38:27.222006 kernel: BTRFS info (device dm-0): first mount of filesystem a6ae7f96-a076-4d3c-81ed-46dd341492f8 Dec 12 18:38:27.222042 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Dec 12 18:38:27.236973 kernel: BTRFS info (device dm-0): enabling ssd optimizations Dec 12 18:38:27.237009 kernel: BTRFS info (device dm-0): disabling log replay at mount time Dec 12 18:38:27.237023 kernel: BTRFS info (device dm-0): enabling free space tree Dec 12 18:38:27.241205 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Dec 12 18:38:27.242439 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Dec 12 18:38:27.243619 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Dec 12 18:38:27.244381 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Dec 12 18:38:27.247430 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Dec 12 18:38:27.275368 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (690) Dec 12 18:38:27.279905 kernel: BTRFS info (device sda6): first mount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:38:27.279939 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 12 18:38:27.291727 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 12 18:38:27.291759 kernel: BTRFS info (device sda6): turning on async discard Dec 12 18:38:27.291773 kernel: BTRFS info (device sda6): enabling free space tree Dec 12 18:38:27.301356 kernel: BTRFS info (device sda6): last unmount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:38:27.302329 systemd[1]: Finished ignition-setup.service - Ignition (setup). Dec 12 18:38:27.304693 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Dec 12 18:38:27.361705 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 12 18:38:27.375182 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 12 18:38:27.417161 systemd-networkd[822]: lo: Link UP Dec 12 18:38:27.417174 systemd-networkd[822]: lo: Gained carrier Dec 12 18:38:27.419880 ignition[765]: Ignition 2.22.0 Dec 12 18:38:27.418708 systemd-networkd[822]: Enumeration completed Dec 12 18:38:27.419886 ignition[765]: Stage: fetch-offline Dec 12 18:38:27.418777 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 12 18:38:27.419909 ignition[765]: no configs at "/usr/lib/ignition/base.d" Dec 12 18:38:27.419705 systemd-networkd[822]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:27.419915 ignition[765]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 12 18:38:27.419708 systemd-networkd[822]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 12 18:38:27.419981 ignition[765]: parsed url from cmdline: "" Dec 12 18:38:27.420138 systemd[1]: Reached target network.target - Network. Dec 12 18:38:27.419983 ignition[765]: no config URL provided Dec 12 18:38:27.420891 systemd-networkd[822]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:27.419987 ignition[765]: reading system config file "/usr/lib/ignition/user.ign" Dec 12 18:38:27.420894 systemd-networkd[822]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 12 18:38:27.419992 ignition[765]: no config at "/usr/lib/ignition/user.ign" Dec 12 18:38:27.421915 systemd-networkd[822]: eth0: Link UP Dec 12 18:38:27.419996 ignition[765]: failed to fetch config: resource requires networking Dec 12 18:38:27.422065 systemd-networkd[822]: eth1: Link UP Dec 12 18:38:27.420104 ignition[765]: Ignition finished successfully Dec 12 18:38:27.422192 systemd-networkd[822]: eth0: Gained carrier Dec 12 18:38:27.422200 systemd-networkd[822]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:27.423710 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Dec 12 18:38:27.426603 systemd-networkd[822]: eth1: Gained carrier Dec 12 18:38:27.426613 systemd-networkd[822]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:27.427827 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Dec 12 18:38:27.453254 ignition[830]: Ignition 2.22.0 Dec 12 18:38:27.453265 ignition[830]: Stage: fetch Dec 12 18:38:27.453403 ignition[830]: no configs at "/usr/lib/ignition/base.d" Dec 12 18:38:27.460424 systemd-networkd[822]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Dec 12 18:38:27.453411 ignition[830]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 12 18:38:27.453471 ignition[830]: parsed url from cmdline: "" Dec 12 18:38:27.453474 ignition[830]: no config URL provided Dec 12 18:38:27.453479 ignition[830]: reading system config file "/usr/lib/ignition/user.ign" Dec 12 18:38:27.453484 ignition[830]: no config at "/usr/lib/ignition/user.ign" Dec 12 18:38:27.453509 ignition[830]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Dec 12 18:38:27.453629 ignition[830]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Dec 12 18:38:27.476381 systemd-networkd[822]: eth0: DHCPv4 address 65.21.248.196/32, gateway 172.31.1.1 acquired from 172.31.1.1 Dec 12 18:38:27.653935 ignition[830]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Dec 12 18:38:27.659422 ignition[830]: GET result: OK Dec 12 18:38:27.659553 ignition[830]: parsing config with SHA512: dce9401ac4321efd68f1adad891fd098ca4b130ff9ac8a6e126cd490e64f11ee95f35ca09d0651dfcd16c6fa12f136e34dbf5113e39fd0ce6b746c59937f2355 Dec 12 18:38:27.667413 unknown[830]: fetched base config from "system" Dec 12 18:38:27.667431 unknown[830]: fetched base config from "system" Dec 12 18:38:27.668083 ignition[830]: fetch: fetch complete Dec 12 18:38:27.667444 unknown[830]: fetched user config from "hetzner" Dec 12 18:38:27.668091 ignition[830]: fetch: fetch passed Dec 12 18:38:27.671195 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Dec 12 18:38:27.668159 ignition[830]: Ignition finished successfully Dec 12 18:38:27.675925 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Dec 12 18:38:27.709151 ignition[837]: Ignition 2.22.0 Dec 12 18:38:27.709166 ignition[837]: Stage: kargs Dec 12 18:38:27.710534 ignition[837]: no configs at "/usr/lib/ignition/base.d" Dec 12 18:38:27.710547 ignition[837]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 12 18:38:27.713109 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Dec 12 18:38:27.711597 ignition[837]: kargs: kargs passed Dec 12 18:38:27.715774 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Dec 12 18:38:27.711641 ignition[837]: Ignition finished successfully Dec 12 18:38:27.743604 ignition[844]: Ignition 2.22.0 Dec 12 18:38:27.743619 ignition[844]: Stage: disks Dec 12 18:38:27.743741 ignition[844]: no configs at "/usr/lib/ignition/base.d" Dec 12 18:38:27.745644 systemd[1]: Finished ignition-disks.service - Ignition (disks). Dec 12 18:38:27.743750 ignition[844]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 12 18:38:27.747627 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Dec 12 18:38:27.744501 ignition[844]: disks: disks passed Dec 12 18:38:27.749509 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 12 18:38:27.744541 ignition[844]: Ignition finished successfully Dec 12 18:38:27.751763 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 12 18:38:27.754127 systemd[1]: Reached target sysinit.target - System Initialization. Dec 12 18:38:27.756693 systemd[1]: Reached target basic.target - Basic System. Dec 12 18:38:27.759500 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Dec 12 18:38:27.780044 systemd-fsck[853]: ROOT: clean, 15/1628000 files, 120826/1617920 blocks Dec 12 18:38:27.782595 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Dec 12 18:38:27.784802 systemd[1]: Mounting sysroot.mount - /sysroot... Dec 12 18:38:27.894379 kernel: EXT4-fs (sda9): mounted filesystem e48ca59c-1206-4abd-b121-5e9b35e49852 r/w with ordered data mode. Quota mode: none. Dec 12 18:38:27.895404 systemd[1]: Mounted sysroot.mount - /sysroot. Dec 12 18:38:27.896610 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Dec 12 18:38:27.899598 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 12 18:38:27.903417 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Dec 12 18:38:27.906579 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Dec 12 18:38:27.910431 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 12 18:38:27.911856 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Dec 12 18:38:27.918716 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Dec 12 18:38:27.921710 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Dec 12 18:38:27.938033 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (861) Dec 12 18:38:27.938076 kernel: BTRFS info (device sda6): first mount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:38:27.946738 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 12 18:38:27.975717 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 12 18:38:27.975827 kernel: BTRFS info (device sda6): turning on async discard Dec 12 18:38:27.975873 kernel: BTRFS info (device sda6): enabling free space tree Dec 12 18:38:27.981437 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 12 18:38:27.984753 coreos-metadata[863]: Dec 12 18:38:27.984 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Dec 12 18:38:27.986585 coreos-metadata[863]: Dec 12 18:38:27.986 INFO Fetch successful Dec 12 18:38:27.987954 coreos-metadata[863]: Dec 12 18:38:27.987 INFO wrote hostname ci-4459-2-2-7-adcd881523 to /sysroot/etc/hostname Dec 12 18:38:27.990778 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 12 18:38:27.995527 initrd-setup-root[889]: cut: /sysroot/etc/passwd: No such file or directory Dec 12 18:38:28.000153 initrd-setup-root[896]: cut: /sysroot/etc/group: No such file or directory Dec 12 18:38:28.003872 initrd-setup-root[903]: cut: /sysroot/etc/shadow: No such file or directory Dec 12 18:38:28.008085 initrd-setup-root[910]: cut: /sysroot/etc/gshadow: No such file or directory Dec 12 18:38:28.091884 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Dec 12 18:38:28.095429 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Dec 12 18:38:28.099462 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Dec 12 18:38:28.111370 kernel: BTRFS info (device sda6): last unmount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:38:28.126024 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Dec 12 18:38:28.139429 ignition[979]: INFO : Ignition 2.22.0 Dec 12 18:38:28.139429 ignition[979]: INFO : Stage: mount Dec 12 18:38:28.142150 ignition[979]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 12 18:38:28.142150 ignition[979]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 12 18:38:28.142150 ignition[979]: INFO : mount: mount passed Dec 12 18:38:28.142150 ignition[979]: INFO : Ignition finished successfully Dec 12 18:38:28.141689 systemd[1]: Finished ignition-mount.service - Ignition (mount). Dec 12 18:38:28.144434 systemd[1]: Starting ignition-files.service - Ignition (files)... Dec 12 18:38:28.214515 systemd[1]: sysroot-oem.mount: Deactivated successfully. Dec 12 18:38:28.216634 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 12 18:38:28.243377 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (989) Dec 12 18:38:28.248013 kernel: BTRFS info (device sda6): first mount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:38:28.248049 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 12 18:38:28.258844 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 12 18:38:28.258911 kernel: BTRFS info (device sda6): turning on async discard Dec 12 18:38:28.258934 kernel: BTRFS info (device sda6): enabling free space tree Dec 12 18:38:28.263552 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 12 18:38:28.298863 ignition[1006]: INFO : Ignition 2.22.0 Dec 12 18:38:28.300119 ignition[1006]: INFO : Stage: files Dec 12 18:38:28.300119 ignition[1006]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 12 18:38:28.300119 ignition[1006]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 12 18:38:28.300119 ignition[1006]: DEBUG : files: compiled without relabeling support, skipping Dec 12 18:38:28.305412 ignition[1006]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 12 18:38:28.305412 ignition[1006]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 12 18:38:28.305412 ignition[1006]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 12 18:38:28.305412 ignition[1006]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 12 18:38:28.305412 ignition[1006]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 12 18:38:28.304791 unknown[1006]: wrote ssh authorized keys file for user: core Dec 12 18:38:28.312438 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Dec 12 18:38:28.312438 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Dec 12 18:38:28.442857 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Dec 12 18:38:28.648478 systemd-networkd[822]: eth0: Gained IPv6LL Dec 12 18:38:28.648837 systemd-networkd[822]: eth1: Gained IPv6LL Dec 12 18:38:28.750846 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Dec 12 18:38:28.750846 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Dec 12 18:38:28.754226 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Dec 12 18:38:28.754226 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Dec 12 18:38:28.754226 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Dec 12 18:38:28.754226 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 12 18:38:28.754226 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 12 18:38:28.754226 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 12 18:38:28.754226 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 12 18:38:28.762996 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 12 18:38:28.762996 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 12 18:38:28.762996 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Dec 12 18:38:28.762996 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Dec 12 18:38:28.762996 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Dec 12 18:38:28.762996 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Dec 12 18:38:29.333023 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Dec 12 18:38:30.433702 ignition[1006]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Dec 12 18:38:30.433702 ignition[1006]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Dec 12 18:38:30.436772 ignition[1006]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 12 18:38:30.438115 ignition[1006]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 12 18:38:30.438115 ignition[1006]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Dec 12 18:38:30.438115 ignition[1006]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Dec 12 18:38:30.438115 ignition[1006]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Dec 12 18:38:30.438115 ignition[1006]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Dec 12 18:38:30.438115 ignition[1006]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Dec 12 18:38:30.438115 ignition[1006]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Dec 12 18:38:30.438115 ignition[1006]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Dec 12 18:38:30.455515 ignition[1006]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 12 18:38:30.455515 ignition[1006]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 12 18:38:30.455515 ignition[1006]: INFO : files: files passed Dec 12 18:38:30.455515 ignition[1006]: INFO : Ignition finished successfully Dec 12 18:38:30.441026 systemd[1]: Finished ignition-files.service - Ignition (files). Dec 12 18:38:30.447447 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Dec 12 18:38:30.450267 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Dec 12 18:38:30.461736 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 12 18:38:30.461815 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Dec 12 18:38:30.468781 initrd-setup-root-after-ignition[1035]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 12 18:38:30.469984 initrd-setup-root-after-ignition[1035]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Dec 12 18:38:30.471164 initrd-setup-root-after-ignition[1039]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 12 18:38:30.470840 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 12 18:38:30.472265 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Dec 12 18:38:30.474546 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Dec 12 18:38:30.516715 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 12 18:38:30.516809 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Dec 12 18:38:30.518402 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Dec 12 18:38:30.519756 systemd[1]: Reached target initrd.target - Initrd Default Target. Dec 12 18:38:30.521248 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Dec 12 18:38:30.521993 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Dec 12 18:38:30.543356 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 12 18:38:30.546675 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Dec 12 18:38:30.570160 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Dec 12 18:38:30.571724 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 12 18:38:30.572526 systemd[1]: Stopped target timers.target - Timer Units. Dec 12 18:38:30.574093 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 12 18:38:30.574221 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 12 18:38:30.575979 systemd[1]: Stopped target initrd.target - Initrd Default Target. Dec 12 18:38:30.577018 systemd[1]: Stopped target basic.target - Basic System. Dec 12 18:38:30.578453 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Dec 12 18:38:30.579859 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Dec 12 18:38:30.581198 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Dec 12 18:38:30.582776 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Dec 12 18:38:30.584254 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Dec 12 18:38:30.585831 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Dec 12 18:38:30.587376 systemd[1]: Stopped target sysinit.target - System Initialization. Dec 12 18:38:30.588886 systemd[1]: Stopped target local-fs.target - Local File Systems. Dec 12 18:38:30.590435 systemd[1]: Stopped target swap.target - Swaps. Dec 12 18:38:30.592151 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 12 18:38:30.592387 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Dec 12 18:38:30.594491 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Dec 12 18:38:30.595785 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 12 18:38:30.597491 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Dec 12 18:38:30.597915 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 12 18:38:30.599425 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 12 18:38:30.599628 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Dec 12 18:38:30.602059 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 12 18:38:30.602264 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 12 18:38:30.603496 systemd[1]: ignition-files.service: Deactivated successfully. Dec 12 18:38:30.603696 systemd[1]: Stopped ignition-files.service - Ignition (files). Dec 12 18:38:30.612701 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Dec 12 18:38:30.612899 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 12 18:38:30.616582 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Dec 12 18:38:30.621703 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Dec 12 18:38:30.625533 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 12 18:38:30.625698 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Dec 12 18:38:30.628485 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 12 18:38:30.628692 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Dec 12 18:38:30.638004 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 12 18:38:30.638154 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Dec 12 18:38:30.655412 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 12 18:38:30.661052 ignition[1059]: INFO : Ignition 2.22.0 Dec 12 18:38:30.661052 ignition[1059]: INFO : Stage: umount Dec 12 18:38:30.664252 ignition[1059]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 12 18:38:30.664252 ignition[1059]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 12 18:38:30.664252 ignition[1059]: INFO : umount: umount passed Dec 12 18:38:30.664252 ignition[1059]: INFO : Ignition finished successfully Dec 12 18:38:30.665634 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 12 18:38:30.666755 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Dec 12 18:38:30.669804 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 12 18:38:30.669856 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Dec 12 18:38:30.673031 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 12 18:38:30.673086 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Dec 12 18:38:30.675398 systemd[1]: ignition-fetch.service: Deactivated successfully. Dec 12 18:38:30.675451 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Dec 12 18:38:30.678820 systemd[1]: Stopped target network.target - Network. Dec 12 18:38:30.682472 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 12 18:38:30.682523 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Dec 12 18:38:30.684021 systemd[1]: Stopped target paths.target - Path Units. Dec 12 18:38:30.686635 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 12 18:38:30.691422 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 12 18:38:30.692743 systemd[1]: Stopped target slices.target - Slice Units. Dec 12 18:38:30.694221 systemd[1]: Stopped target sockets.target - Socket Units. Dec 12 18:38:30.695843 systemd[1]: iscsid.socket: Deactivated successfully. Dec 12 18:38:30.695888 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Dec 12 18:38:30.697363 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 12 18:38:30.697392 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 12 18:38:30.698673 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 12 18:38:30.698716 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Dec 12 18:38:30.699972 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Dec 12 18:38:30.700005 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Dec 12 18:38:30.701416 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Dec 12 18:38:30.702750 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Dec 12 18:38:30.704331 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 12 18:38:30.704435 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Dec 12 18:38:30.706414 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 12 18:38:30.706475 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Dec 12 18:38:30.708610 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 12 18:38:30.708700 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Dec 12 18:38:30.711677 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Dec 12 18:38:30.712447 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 12 18:38:30.712520 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Dec 12 18:38:30.714579 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Dec 12 18:38:30.714830 systemd[1]: Stopped target network-pre.target - Preparation for Network. Dec 12 18:38:30.715794 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 12 18:38:30.715819 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Dec 12 18:38:30.717724 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Dec 12 18:38:30.719720 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 12 18:38:30.719759 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 12 18:38:30.720432 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 12 18:38:30.720464 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 12 18:38:30.723443 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 12 18:38:30.723476 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Dec 12 18:38:30.724755 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Dec 12 18:38:30.724790 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 12 18:38:30.727330 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 12 18:38:30.729275 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Dec 12 18:38:30.729361 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Dec 12 18:38:30.737777 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 12 18:38:30.737850 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Dec 12 18:38:30.740617 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 12 18:38:30.740725 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 12 18:38:30.741819 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 12 18:38:30.741863 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Dec 12 18:38:30.742973 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 12 18:38:30.742998 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Dec 12 18:38:30.744257 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 12 18:38:30.744326 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Dec 12 18:38:30.746229 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 12 18:38:30.746302 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Dec 12 18:38:30.747682 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 12 18:38:30.747728 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 12 18:38:30.751453 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Dec 12 18:38:30.752707 systemd[1]: systemd-network-generator.service: Deactivated successfully. Dec 12 18:38:30.752758 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Dec 12 18:38:30.755431 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 12 18:38:30.755479 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 12 18:38:30.756535 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Dec 12 18:38:30.756572 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 12 18:38:30.758229 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 12 18:38:30.758260 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Dec 12 18:38:30.765406 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 12 18:38:30.765455 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:38:30.767460 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Dec 12 18:38:30.767504 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Dec 12 18:38:30.767529 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Dec 12 18:38:30.767577 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Dec 12 18:38:30.772016 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 12 18:38:30.772078 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Dec 12 18:38:30.773613 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Dec 12 18:38:30.775630 systemd[1]: Starting initrd-switch-root.service - Switch Root... Dec 12 18:38:30.789399 systemd[1]: Switching root. Dec 12 18:38:30.827032 systemd-journald[198]: Journal stopped Dec 12 18:38:31.793443 systemd-journald[198]: Received SIGTERM from PID 1 (systemd). Dec 12 18:38:31.793489 kernel: SELinux: policy capability network_peer_controls=1 Dec 12 18:38:31.793500 kernel: SELinux: policy capability open_perms=1 Dec 12 18:38:31.793510 kernel: SELinux: policy capability extended_socket_class=1 Dec 12 18:38:31.793517 kernel: SELinux: policy capability always_check_network=0 Dec 12 18:38:31.793525 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 12 18:38:31.793533 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 12 18:38:31.793541 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 12 18:38:31.793549 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 12 18:38:31.793556 kernel: SELinux: policy capability userspace_initial_context=0 Dec 12 18:38:31.793563 kernel: audit: type=1403 audit(1765564710.961:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 12 18:38:31.793572 systemd[1]: Successfully loaded SELinux policy in 70.926ms. Dec 12 18:38:31.793588 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 7.975ms. Dec 12 18:38:31.793597 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Dec 12 18:38:31.793606 systemd[1]: Detected virtualization kvm. Dec 12 18:38:31.793614 systemd[1]: Detected architecture x86-64. Dec 12 18:38:31.793622 systemd[1]: Detected first boot. Dec 12 18:38:31.793630 systemd[1]: Hostname set to . Dec 12 18:38:31.793638 systemd[1]: Initializing machine ID from VM UUID. Dec 12 18:38:31.793647 kernel: Guest personality initialized and is inactive Dec 12 18:38:31.793658 zram_generator::config[1102]: No configuration found. Dec 12 18:38:31.793669 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Dec 12 18:38:31.793677 kernel: Initialized host personality Dec 12 18:38:31.793684 kernel: NET: Registered PF_VSOCK protocol family Dec 12 18:38:31.793692 systemd[1]: Populated /etc with preset unit settings. Dec 12 18:38:31.793700 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Dec 12 18:38:31.793709 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 12 18:38:31.793717 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Dec 12 18:38:31.793729 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 12 18:38:31.793738 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Dec 12 18:38:31.793747 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Dec 12 18:38:31.793756 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Dec 12 18:38:31.793764 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Dec 12 18:38:31.793772 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Dec 12 18:38:31.793782 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Dec 12 18:38:31.793791 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Dec 12 18:38:31.793803 systemd[1]: Created slice user.slice - User and Session Slice. Dec 12 18:38:31.793819 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 12 18:38:31.793835 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 12 18:38:31.793849 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Dec 12 18:38:31.793864 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Dec 12 18:38:31.793880 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Dec 12 18:38:31.793894 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 12 18:38:31.793911 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Dec 12 18:38:31.793927 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 12 18:38:31.793941 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 12 18:38:31.793954 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Dec 12 18:38:31.793971 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Dec 12 18:38:31.793984 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Dec 12 18:38:31.794000 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Dec 12 18:38:31.794013 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 12 18:38:31.794027 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 12 18:38:31.794040 systemd[1]: Reached target slices.target - Slice Units. Dec 12 18:38:31.794053 systemd[1]: Reached target swap.target - Swaps. Dec 12 18:38:31.794066 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Dec 12 18:38:31.794079 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Dec 12 18:38:31.794093 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Dec 12 18:38:31.794106 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 12 18:38:31.794120 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 12 18:38:31.794133 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 12 18:38:31.794146 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Dec 12 18:38:31.794159 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Dec 12 18:38:31.794172 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Dec 12 18:38:31.794185 systemd[1]: Mounting media.mount - External Media Directory... Dec 12 18:38:31.794199 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:38:31.794214 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Dec 12 18:38:31.794230 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Dec 12 18:38:31.794247 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Dec 12 18:38:31.794259 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 12 18:38:31.794268 systemd[1]: Reached target machines.target - Containers. Dec 12 18:38:31.794277 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Dec 12 18:38:31.794286 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 12 18:38:31.794308 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 12 18:38:31.794316 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Dec 12 18:38:31.794324 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 12 18:38:31.794349 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 12 18:38:31.794359 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 12 18:38:31.794368 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Dec 12 18:38:31.794378 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 12 18:38:31.794387 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 12 18:38:31.794395 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 12 18:38:31.794404 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Dec 12 18:38:31.794413 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 12 18:38:31.794421 systemd[1]: Stopped systemd-fsck-usr.service. Dec 12 18:38:31.794431 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 12 18:38:31.794440 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 12 18:38:31.794448 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 12 18:38:31.794458 kernel: loop: module loaded Dec 12 18:38:31.794466 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 12 18:38:31.794477 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Dec 12 18:38:31.794485 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Dec 12 18:38:31.794495 kernel: ACPI: bus type drm_connector registered Dec 12 18:38:31.794502 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 12 18:38:31.794511 systemd[1]: verity-setup.service: Deactivated successfully. Dec 12 18:38:31.794519 kernel: fuse: init (API version 7.41) Dec 12 18:38:31.794527 systemd[1]: Stopped verity-setup.service. Dec 12 18:38:31.794535 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:38:31.794545 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Dec 12 18:38:31.794554 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Dec 12 18:38:31.794562 systemd[1]: Mounted media.mount - External Media Directory. Dec 12 18:38:31.794570 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Dec 12 18:38:31.794578 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Dec 12 18:38:31.794605 systemd-journald[1193]: Collecting audit messages is disabled. Dec 12 18:38:31.794626 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Dec 12 18:38:31.794635 systemd-journald[1193]: Journal started Dec 12 18:38:31.794654 systemd-journald[1193]: Runtime Journal (/run/log/journal/6bc877b412ba488ab1eb8306b72c236b) is 4.7M, max 38.3M, 33.5M free. Dec 12 18:38:31.430587 systemd[1]: Queued start job for default target multi-user.target. Dec 12 18:38:31.442062 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Dec 12 18:38:31.442584 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 12 18:38:31.799412 systemd[1]: Started systemd-journald.service - Journal Service. Dec 12 18:38:31.799993 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Dec 12 18:38:31.800948 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 12 18:38:31.801879 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 12 18:38:31.802000 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Dec 12 18:38:31.802921 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 12 18:38:31.803087 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 12 18:38:31.804028 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 12 18:38:31.804185 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 12 18:38:31.805077 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 12 18:38:31.805182 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 12 18:38:31.806207 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 12 18:38:31.806470 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Dec 12 18:38:31.807304 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 12 18:38:31.807604 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 12 18:38:31.808524 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 12 18:38:31.809543 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 12 18:38:31.810562 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Dec 12 18:38:31.811622 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Dec 12 18:38:31.818850 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 12 18:38:31.821421 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Dec 12 18:38:31.827400 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Dec 12 18:38:31.828046 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 12 18:38:31.828070 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 12 18:38:31.829780 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Dec 12 18:38:31.836656 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Dec 12 18:38:31.837373 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 12 18:38:31.840286 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Dec 12 18:38:31.841788 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Dec 12 18:38:31.842509 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 12 18:38:31.844491 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Dec 12 18:38:31.845138 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 12 18:38:31.849991 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 12 18:38:31.853595 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Dec 12 18:38:31.865701 systemd-journald[1193]: Time spent on flushing to /var/log/journal/6bc877b412ba488ab1eb8306b72c236b is 30.186ms for 1169 entries. Dec 12 18:38:31.865701 systemd-journald[1193]: System Journal (/var/log/journal/6bc877b412ba488ab1eb8306b72c236b) is 8M, max 584.8M, 576.8M free. Dec 12 18:38:31.914820 systemd-journald[1193]: Received client request to flush runtime journal. Dec 12 18:38:31.914859 kernel: loop0: detected capacity change from 0 to 128560 Dec 12 18:38:31.860548 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 12 18:38:31.865379 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Dec 12 18:38:31.870162 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Dec 12 18:38:31.872483 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 12 18:38:31.875268 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Dec 12 18:38:31.877245 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Dec 12 18:38:31.882860 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Dec 12 18:38:31.905744 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 12 18:38:31.916171 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Dec 12 18:38:31.930216 systemd-tmpfiles[1228]: ACLs are not supported, ignoring. Dec 12 18:38:31.930238 systemd-tmpfiles[1228]: ACLs are not supported, ignoring. Dec 12 18:38:31.931487 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Dec 12 18:38:31.935363 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 12 18:38:31.939602 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 12 18:38:31.942429 systemd[1]: Starting systemd-sysusers.service - Create System Users... Dec 12 18:38:31.949614 kernel: loop1: detected capacity change from 0 to 110984 Dec 12 18:38:31.986667 systemd[1]: Finished systemd-sysusers.service - Create System Users. Dec 12 18:38:31.993369 kernel: loop2: detected capacity change from 0 to 8 Dec 12 18:38:31.990525 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 12 18:38:32.006447 kernel: loop3: detected capacity change from 0 to 229808 Dec 12 18:38:32.017431 systemd-tmpfiles[1249]: ACLs are not supported, ignoring. Dec 12 18:38:32.017449 systemd-tmpfiles[1249]: ACLs are not supported, ignoring. Dec 12 18:38:32.020723 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 12 18:38:32.040383 kernel: loop4: detected capacity change from 0 to 128560 Dec 12 18:38:32.070380 kernel: loop5: detected capacity change from 0 to 110984 Dec 12 18:38:32.085380 kernel: loop6: detected capacity change from 0 to 8 Dec 12 18:38:32.090636 kernel: loop7: detected capacity change from 0 to 229808 Dec 12 18:38:32.119808 (sd-merge)[1254]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Dec 12 18:38:32.120854 (sd-merge)[1254]: Merged extensions into '/usr'. Dec 12 18:38:32.129741 systemd[1]: Reload requested from client PID 1227 ('systemd-sysext') (unit systemd-sysext.service)... Dec 12 18:38:32.130152 systemd[1]: Reloading... Dec 12 18:38:32.198549 zram_generator::config[1276]: No configuration found. Dec 12 18:38:32.368331 ldconfig[1222]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 12 18:38:32.393004 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 12 18:38:32.393300 systemd[1]: Reloading finished in 262 ms. Dec 12 18:38:32.423644 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Dec 12 18:38:32.424820 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Dec 12 18:38:32.433472 systemd[1]: Starting ensure-sysext.service... Dec 12 18:38:32.434914 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 12 18:38:32.462944 systemd[1]: Reload requested from client PID 1323 ('systemctl') (unit ensure-sysext.service)... Dec 12 18:38:32.463062 systemd[1]: Reloading... Dec 12 18:38:32.464738 systemd-tmpfiles[1324]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Dec 12 18:38:32.465055 systemd-tmpfiles[1324]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Dec 12 18:38:32.465252 systemd-tmpfiles[1324]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 12 18:38:32.465512 systemd-tmpfiles[1324]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Dec 12 18:38:32.466033 systemd-tmpfiles[1324]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 12 18:38:32.466208 systemd-tmpfiles[1324]: ACLs are not supported, ignoring. Dec 12 18:38:32.466251 systemd-tmpfiles[1324]: ACLs are not supported, ignoring. Dec 12 18:38:32.473150 systemd-tmpfiles[1324]: Detected autofs mount point /boot during canonicalization of boot. Dec 12 18:38:32.473157 systemd-tmpfiles[1324]: Skipping /boot Dec 12 18:38:32.479238 systemd-tmpfiles[1324]: Detected autofs mount point /boot during canonicalization of boot. Dec 12 18:38:32.479250 systemd-tmpfiles[1324]: Skipping /boot Dec 12 18:38:32.508387 zram_generator::config[1350]: No configuration found. Dec 12 18:38:32.642872 systemd[1]: Reloading finished in 179 ms. Dec 12 18:38:32.660557 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Dec 12 18:38:32.665141 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 12 18:38:32.669203 systemd[1]: Starting audit-rules.service - Load Audit Rules... Dec 12 18:38:32.672555 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Dec 12 18:38:32.677119 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Dec 12 18:38:32.681434 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 12 18:38:32.684764 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 12 18:38:32.689306 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Dec 12 18:38:32.702622 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Dec 12 18:38:32.706897 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:38:32.707029 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 12 18:38:32.708194 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 12 18:38:32.715603 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 12 18:38:32.719016 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 12 18:38:32.720468 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 12 18:38:32.720565 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 12 18:38:32.720641 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:38:32.726002 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:38:32.726156 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 12 18:38:32.726726 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 12 18:38:32.726842 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 12 18:38:32.726945 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:38:32.729017 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Dec 12 18:38:32.735644 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:38:32.735842 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 12 18:38:32.737989 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 12 18:38:32.739461 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 12 18:38:32.739550 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 12 18:38:32.741839 systemd[1]: Starting systemd-update-done.service - Update is Completed... Dec 12 18:38:32.742525 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:38:32.744393 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Dec 12 18:38:32.748933 systemd[1]: Finished ensure-sysext.service. Dec 12 18:38:32.749613 systemd-udevd[1400]: Using default interface naming scheme 'v255'. Dec 12 18:38:32.756756 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Dec 12 18:38:32.768190 systemd[1]: Finished systemd-update-done.service - Update is Completed. Dec 12 18:38:32.771261 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 12 18:38:32.771459 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 12 18:38:32.772566 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 12 18:38:32.773059 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 12 18:38:32.775027 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 12 18:38:32.775167 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 12 18:38:32.778609 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 12 18:38:32.779129 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 12 18:38:32.781426 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 12 18:38:32.781474 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 12 18:38:32.790582 augenrules[1436]: No rules Dec 12 18:38:32.791280 systemd[1]: audit-rules.service: Deactivated successfully. Dec 12 18:38:32.791500 systemd[1]: Finished audit-rules.service - Load Audit Rules. Dec 12 18:38:32.795808 systemd[1]: Started systemd-userdbd.service - User Database Manager. Dec 12 18:38:32.806930 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 12 18:38:32.809359 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 12 18:38:32.813404 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Dec 12 18:38:32.814207 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 12 18:38:32.897532 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Dec 12 18:38:32.941205 kernel: mousedev: PS/2 mouse device common for all mice Dec 12 18:38:32.961440 systemd-networkd[1447]: lo: Link UP Dec 12 18:38:32.961446 systemd-networkd[1447]: lo: Gained carrier Dec 12 18:38:32.962910 systemd-networkd[1447]: Enumeration completed Dec 12 18:38:32.962983 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 12 18:38:32.968479 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Dec 12 18:38:32.977477 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Dec 12 18:38:32.980616 systemd-networkd[1447]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:32.980619 systemd-networkd[1447]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 12 18:38:32.981121 systemd-networkd[1447]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:32.981124 systemd-networkd[1447]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 12 18:38:32.981517 systemd-networkd[1447]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:32.981541 systemd-networkd[1447]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:32.981558 systemd-networkd[1447]: eth0: Link UP Dec 12 18:38:32.981657 systemd-networkd[1447]: eth0: Gained carrier Dec 12 18:38:32.981667 systemd-networkd[1447]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:32.987540 systemd-networkd[1447]: eth1: Link UP Dec 12 18:38:32.988151 systemd-networkd[1447]: eth1: Gained carrier Dec 12 18:38:32.988361 systemd-networkd[1447]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:38:33.005083 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Dec 12 18:38:33.023100 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Dec 12 18:38:33.024464 systemd[1]: Reached target time-set.target - System Time Set. Dec 12 18:38:33.030039 systemd-resolved[1399]: Positive Trust Anchors: Dec 12 18:38:33.030054 systemd-resolved[1399]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 12 18:38:33.030079 systemd-resolved[1399]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 12 18:38:33.031412 systemd-networkd[1447]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Dec 12 18:38:33.032824 systemd-timesyncd[1426]: Network configuration changed, trying to establish connection. Dec 12 18:38:33.033523 systemd-resolved[1399]: Using system hostname 'ci-4459-2-2-7-adcd881523'. Dec 12 18:38:33.034714 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 12 18:38:33.035779 systemd[1]: Reached target network.target - Network. Dec 12 18:38:33.036825 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 12 18:38:33.037964 systemd[1]: Reached target sysinit.target - System Initialization. Dec 12 18:38:33.039279 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Dec 12 18:38:33.039617 systemd-networkd[1447]: eth0: DHCPv4 address 65.21.248.196/32, gateway 172.31.1.1 acquired from 172.31.1.1 Dec 12 18:38:33.040310 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Dec 12 18:38:33.041390 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Dec 12 18:38:33.042733 systemd[1]: Started logrotate.timer - Daily rotation of log files. Dec 12 18:38:33.044260 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Dec 12 18:38:33.045352 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Dec 12 18:38:33.047241 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 12 18:38:33.047277 systemd[1]: Reached target paths.target - Path Units. Dec 12 18:38:33.048286 systemd[1]: Reached target timers.target - Timer Units. Dec 12 18:38:33.050679 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Dec 12 18:38:33.053182 systemd[1]: Starting docker.socket - Docker Socket for the API... Dec 12 18:38:33.057920 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Dec 12 18:38:33.059445 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Dec 12 18:38:33.060482 systemd[1]: Reached target ssh-access.target - SSH Access Available. Dec 12 18:38:33.065823 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Dec 12 18:38:33.067708 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Dec 12 18:38:33.069212 systemd[1]: Listening on docker.socket - Docker Socket for the API. Dec 12 18:38:33.078439 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input4 Dec 12 18:38:33.083852 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Dec 12 18:38:33.084557 systemd[1]: Reached target sockets.target - Socket Units. Dec 12 18:38:33.085146 systemd[1]: Reached target basic.target - Basic System. Dec 12 18:38:33.085818 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Dec 12 18:38:33.085858 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Dec 12 18:38:33.086571 systemd[1]: Starting containerd.service - containerd container runtime... Dec 12 18:38:33.089454 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Dec 12 18:38:33.093644 kernel: ACPI: button: Power Button [PWRF] Dec 12 18:38:33.092439 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Dec 12 18:38:33.096472 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Dec 12 18:38:33.102953 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Dec 12 18:38:33.107786 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Dec 12 18:38:33.108921 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Dec 12 18:38:33.110687 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Dec 12 18:38:33.116747 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Dec 12 18:38:33.122217 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Dec 12 18:38:33.123565 extend-filesystems[1506]: Found /dev/sda6 Dec 12 18:38:33.132433 extend-filesystems[1506]: Found /dev/sda9 Dec 12 18:38:33.132433 extend-filesystems[1506]: Checking size of /dev/sda9 Dec 12 18:38:33.124675 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Dec 12 18:38:33.139033 oslogin_cache_refresh[1507]: Refreshing passwd entry cache Dec 12 18:38:33.139807 google_oslogin_nss_cache[1507]: oslogin_cache_refresh[1507]: Refreshing passwd entry cache Dec 12 18:38:33.128051 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Dec 12 18:38:33.130801 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Dec 12 18:38:33.140786 systemd[1]: Starting systemd-logind.service - User Login Management... Dec 12 18:38:33.146254 google_oslogin_nss_cache[1507]: oslogin_cache_refresh[1507]: Failure getting users, quitting Dec 12 18:38:33.146254 google_oslogin_nss_cache[1507]: oslogin_cache_refresh[1507]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Dec 12 18:38:33.146254 google_oslogin_nss_cache[1507]: oslogin_cache_refresh[1507]: Refreshing group entry cache Dec 12 18:38:33.146254 google_oslogin_nss_cache[1507]: oslogin_cache_refresh[1507]: Failure getting groups, quitting Dec 12 18:38:33.146254 google_oslogin_nss_cache[1507]: oslogin_cache_refresh[1507]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Dec 12 18:38:33.146676 extend-filesystems[1506]: Resized partition /dev/sda9 Dec 12 18:38:33.147668 jq[1503]: false Dec 12 18:38:33.144460 oslogin_cache_refresh[1507]: Failure getting users, quitting Dec 12 18:38:33.143085 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 12 18:38:33.144472 oslogin_cache_refresh[1507]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Dec 12 18:38:33.144502 oslogin_cache_refresh[1507]: Refreshing group entry cache Dec 12 18:38:33.145099 oslogin_cache_refresh[1507]: Failure getting groups, quitting Dec 12 18:38:33.145106 oslogin_cache_refresh[1507]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Dec 12 18:38:33.154375 extend-filesystems[1523]: resize2fs 1.47.3 (8-Jul-2025) Dec 12 18:38:33.151545 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 12 18:38:33.153428 systemd[1]: Starting update-engine.service - Update Engine... Dec 12 18:38:33.162903 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Dec 12 18:38:33.157909 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Dec 12 18:38:33.161646 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Dec 12 18:38:33.163068 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 12 18:38:33.163253 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Dec 12 18:38:33.163501 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Dec 12 18:38:33.163626 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Dec 12 18:38:33.167485 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 12 18:38:33.167625 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Dec 12 18:38:33.170523 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Dec 12 18:38:33.178496 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Dec 12 18:38:33.180756 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Dec 12 18:38:33.183090 jq[1525]: true Dec 12 18:38:33.199140 coreos-metadata[1500]: Dec 12 18:38:33.199 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Dec 12 18:38:33.204264 coreos-metadata[1500]: Dec 12 18:38:33.203 INFO Fetch successful Dec 12 18:38:33.204264 coreos-metadata[1500]: Dec 12 18:38:33.203 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Dec 12 18:38:33.215031 tar[1531]: linux-amd64/LICENSE Dec 12 18:38:33.215753 (ntainerd)[1549]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Dec 12 18:38:33.218220 coreos-metadata[1500]: Dec 12 18:38:33.217 INFO Fetch successful Dec 12 18:38:33.218276 tar[1531]: linux-amd64/helm Dec 12 18:38:33.223645 jq[1542]: true Dec 12 18:38:33.226905 systemd[1]: motdgen.service: Deactivated successfully. Dec 12 18:38:33.227388 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Dec 12 18:38:33.238234 update_engine[1524]: I20251212 18:38:33.235901 1524 main.cc:92] Flatcar Update Engine starting Dec 12 18:38:34.114992 systemd-resolved[1399]: Clock change detected. Flushing caches. Dec 12 18:38:34.115094 systemd-timesyncd[1426]: Contacted time server 162.159.200.1:123 (3.flatcar.pool.ntp.org). Dec 12 18:38:34.115128 systemd-timesyncd[1426]: Initial clock synchronization to Fri 2025-12-12 18:38:34.114950 UTC. Dec 12 18:38:34.126171 systemd-logind[1520]: New seat seat0. Dec 12 18:38:34.126896 systemd[1]: Started systemd-logind.service - User Login Management. Dec 12 18:38:34.131934 dbus-daemon[1501]: [system] SELinux support is enabled Dec 12 18:38:34.132048 systemd[1]: Started dbus.service - D-Bus System Message Bus. Dec 12 18:38:34.137373 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 12 18:38:34.137401 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Dec 12 18:38:34.138068 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 12 18:38:34.138081 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Dec 12 18:38:34.146329 update_engine[1524]: I20251212 18:38:34.145931 1524 update_check_scheduler.cc:74] Next update check in 5m55s Dec 12 18:38:34.149571 systemd[1]: Started update-engine.service - Update Engine. Dec 12 18:38:34.150702 dbus-daemon[1501]: [system] Successfully activated service 'org.freedesktop.systemd1' Dec 12 18:38:34.158862 systemd[1]: Started locksmithd.service - Cluster reboot manager. Dec 12 18:38:34.203848 kernel: EDAC MC: Ver: 3.0.0 Dec 12 18:38:34.228234 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Dec 12 18:38:34.230406 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Dec 12 18:38:34.245538 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Dec 12 18:38:34.250072 extend-filesystems[1523]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Dec 12 18:38:34.250072 extend-filesystems[1523]: old_desc_blocks = 1, new_desc_blocks = 5 Dec 12 18:38:34.250072 extend-filesystems[1523]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Dec 12 18:38:34.249244 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 12 18:38:34.253516 extend-filesystems[1506]: Resized filesystem in /dev/sda9 Dec 12 18:38:34.249478 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Dec 12 18:38:34.252790 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Dec 12 18:38:34.255592 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Dec 12 18:38:34.266319 bash[1582]: Updated "/home/core/.ssh/authorized_keys" Dec 12 18:38:34.266922 sshd_keygen[1548]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 12 18:38:34.267356 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Dec 12 18:38:34.286359 systemd[1]: Starting sshkeys.service... Dec 12 18:38:34.330036 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Dec 12 18:38:34.334424 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Dec 12 18:38:34.358526 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Dec 12 18:38:34.363507 systemd[1]: Starting issuegen.service - Generate /run/issue... Dec 12 18:38:34.371490 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:38:34.384619 systemd[1]: issuegen.service: Deactivated successfully. Dec 12 18:38:34.385236 systemd[1]: Finished issuegen.service - Generate /run/issue. Dec 12 18:38:34.389524 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Dec 12 18:38:34.398218 kernel: [drm] pci: virtio-vga detected at 0000:00:01.0 Dec 12 18:38:34.409466 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Dec 12 18:38:34.412120 systemd[1]: Started getty@tty1.service - Getty on tty1. Dec 12 18:38:34.416352 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Dec 12 18:38:34.420701 coreos-metadata[1607]: Dec 12 18:38:34.420 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Dec 12 18:38:34.417443 systemd[1]: Reached target getty.target - Login Prompts. Dec 12 18:38:34.423519 coreos-metadata[1607]: Dec 12 18:38:34.423 INFO Fetch successful Dec 12 18:38:34.426713 unknown[1607]: wrote ssh authorized keys file for user: core Dec 12 18:38:34.436068 locksmithd[1558]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 12 18:38:34.439222 kernel: virtio-pci 0000:00:01.0: vgaarb: deactivate vga console Dec 12 18:38:34.449468 systemd-logind[1520]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Dec 12 18:38:34.452568 kernel: Console: switching to colour dummy device 80x25 Dec 12 18:38:34.453209 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Dec 12 18:38:34.453243 kernel: [drm] features: -context_init Dec 12 18:38:34.456204 kernel: [drm] number of scanouts: 1 Dec 12 18:38:34.458211 kernel: [drm] number of cap sets: 0 Dec 12 18:38:34.461205 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:01.0 on minor 0 Dec 12 18:38:34.470715 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Dec 12 18:38:34.470756 kernel: Console: switching to colour frame buffer device 160x50 Dec 12 18:38:34.478223 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Dec 12 18:38:34.482519 update-ssh-keys[1630]: Updated "/home/core/.ssh/authorized_keys" Dec 12 18:38:34.484527 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Dec 12 18:38:34.490352 systemd[1]: Finished sshkeys.service. Dec 12 18:38:34.495113 systemd-logind[1520]: Watching system buttons on /dev/input/event3 (Power Button) Dec 12 18:38:34.504099 containerd[1549]: time="2025-12-12T18:38:34Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Dec 12 18:38:34.506126 containerd[1549]: time="2025-12-12T18:38:34.506099929Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Dec 12 18:38:34.525129 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:38:34.538850 containerd[1549]: time="2025-12-12T18:38:34.537172853Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="11.081µs" Dec 12 18:38:34.538850 containerd[1549]: time="2025-12-12T18:38:34.538845359Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Dec 12 18:38:34.538946 containerd[1549]: time="2025-12-12T18:38:34.538865187Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Dec 12 18:38:34.539003 containerd[1549]: time="2025-12-12T18:38:34.538982116Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Dec 12 18:38:34.539025 containerd[1549]: time="2025-12-12T18:38:34.539002334Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Dec 12 18:38:34.539039 containerd[1549]: time="2025-12-12T18:38:34.539023724Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539093 containerd[1549]: time="2025-12-12T18:38:34.539073888Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539093 containerd[1549]: time="2025-12-12T18:38:34.539088676Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539373 containerd[1549]: time="2025-12-12T18:38:34.539347912Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539373 containerd[1549]: time="2025-12-12T18:38:34.539370344Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539427 containerd[1549]: time="2025-12-12T18:38:34.539381175Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539427 containerd[1549]: time="2025-12-12T18:38:34.539387466Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539456 containerd[1549]: time="2025-12-12T18:38:34.539447398Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539619 containerd[1549]: time="2025-12-12T18:38:34.539592611Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539650 containerd[1549]: time="2025-12-12T18:38:34.539629811Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Dec 12 18:38:34.539650 containerd[1549]: time="2025-12-12T18:38:34.539638837Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Dec 12 18:38:34.541056 containerd[1549]: time="2025-12-12T18:38:34.541031378Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Dec 12 18:38:34.541736 containerd[1549]: time="2025-12-12T18:38:34.541711925Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Dec 12 18:38:34.541787 containerd[1549]: time="2025-12-12T18:38:34.541770335Z" level=info msg="metadata content store policy set" policy=shared Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.544939086Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.544979312Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.544992376Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545002184Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545012093Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545020228Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545030187Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545047379Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545055865Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545064130Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545071624Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545083066Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545180459Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Dec 12 18:38:34.547126 containerd[1549]: time="2025-12-12T18:38:34.545270027Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545294663Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545313138Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545330700Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545357009Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545377438Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545396193Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545411612Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545425318Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545439705Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545492013Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545510297Z" level=info msg="Start snapshots syncer" Dec 12 18:38:34.547381 containerd[1549]: time="2025-12-12T18:38:34.545543639Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Dec 12 18:38:34.547546 containerd[1549]: time="2025-12-12T18:38:34.545858681Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Dec 12 18:38:34.547546 containerd[1549]: time="2025-12-12T18:38:34.545902432Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.545941365Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546041593Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546069626Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546085014Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546098760Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546116935Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546137192Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546162871Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546181986Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546209539Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546221371Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546242210Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546251948Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Dec 12 18:38:34.547642 containerd[1549]: time="2025-12-12T18:38:34.546258570Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546266234Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546272035Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546279048Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546291101Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546302863Z" level=info msg="runtime interface created" Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546306650Z" level=info msg="created NRI interface" Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546312521Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546320757Z" level=info msg="Connect containerd service" Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546335885Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Dec 12 18:38:34.547828 containerd[1549]: time="2025-12-12T18:38:34.546863945Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 12 18:38:34.645600 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 12 18:38:34.645914 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:38:34.647803 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:38:34.651140 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:38:34.653610 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Dec 12 18:38:34.692824 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 12 18:38:34.692985 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:38:34.698231 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:38:34.699037 containerd[1549]: time="2025-12-12T18:38:34.699007880Z" level=info msg="Start subscribing containerd event" Dec 12 18:38:34.699089 containerd[1549]: time="2025-12-12T18:38:34.699051613Z" level=info msg="Start recovering state" Dec 12 18:38:34.699195 containerd[1549]: time="2025-12-12T18:38:34.699170896Z" level=info msg="Start event monitor" Dec 12 18:38:34.699225 containerd[1549]: time="2025-12-12T18:38:34.699208116Z" level=info msg="Start cni network conf syncer for default" Dec 12 18:38:34.699225 containerd[1549]: time="2025-12-12T18:38:34.699215089Z" level=info msg="Start streaming server" Dec 12 18:38:34.699257 containerd[1549]: time="2025-12-12T18:38:34.699230248Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Dec 12 18:38:34.699257 containerd[1549]: time="2025-12-12T18:38:34.699235537Z" level=info msg="runtime interface starting up..." Dec 12 18:38:34.699257 containerd[1549]: time="2025-12-12T18:38:34.699240627Z" level=info msg="starting plugins..." Dec 12 18:38:34.699257 containerd[1549]: time="2025-12-12T18:38:34.699252600Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Dec 12 18:38:34.699591 containerd[1549]: time="2025-12-12T18:38:34.699571227Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 12 18:38:34.699740 containerd[1549]: time="2025-12-12T18:38:34.699724334Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 12 18:38:34.699922 systemd[1]: Started containerd.service - containerd container runtime. Dec 12 18:38:34.700276 containerd[1549]: time="2025-12-12T18:38:34.700258977Z" level=info msg="containerd successfully booted in 0.196406s" Dec 12 18:38:34.728563 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:38:34.733178 tar[1531]: linux-amd64/README.md Dec 12 18:38:34.742558 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Dec 12 18:38:35.282370 systemd-networkd[1447]: eth0: Gained IPv6LL Dec 12 18:38:35.284877 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Dec 12 18:38:35.285764 systemd[1]: Reached target network-online.target - Network is Online. Dec 12 18:38:35.288647 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:38:35.290509 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Dec 12 18:38:35.315937 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Dec 12 18:38:35.538387 systemd-networkd[1447]: eth1: Gained IPv6LL Dec 12 18:38:36.358853 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:38:36.359824 systemd[1]: Reached target multi-user.target - Multi-User System. Dec 12 18:38:36.362117 systemd[1]: Startup finished in 3.299s (kernel) + 6.303s (initrd) + 4.597s (userspace) = 14.199s. Dec 12 18:38:36.373665 (kubelet)[1686]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 12 18:38:37.060618 kubelet[1686]: E1212 18:38:37.060530 1686 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 12 18:38:37.065050 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 12 18:38:37.065333 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 12 18:38:37.065791 systemd[1]: kubelet.service: Consumed 1.150s CPU time, 268.6M memory peak. Dec 12 18:38:40.092685 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Dec 12 18:38:40.093711 systemd[1]: Started sshd@0-65.21.248.196:22-147.75.109.163:34826.service - OpenSSH per-connection server daemon (147.75.109.163:34826). Dec 12 18:38:41.080111 sshd[1698]: Accepted publickey for core from 147.75.109.163 port 34826 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:38:41.082609 sshd-session[1698]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:38:41.101585 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Dec 12 18:38:41.102311 systemd-logind[1520]: New session 1 of user core. Dec 12 18:38:41.102453 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Dec 12 18:38:41.126820 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Dec 12 18:38:41.130729 systemd[1]: Starting user@500.service - User Manager for UID 500... Dec 12 18:38:41.140905 (systemd)[1703]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 12 18:38:41.143848 systemd-logind[1520]: New session c1 of user core. Dec 12 18:38:41.283241 systemd[1703]: Queued start job for default target default.target. Dec 12 18:38:41.292969 systemd[1703]: Created slice app.slice - User Application Slice. Dec 12 18:38:41.292993 systemd[1703]: Reached target paths.target - Paths. Dec 12 18:38:41.293024 systemd[1703]: Reached target timers.target - Timers. Dec 12 18:38:41.294137 systemd[1703]: Starting dbus.socket - D-Bus User Message Bus Socket... Dec 12 18:38:41.309867 systemd[1703]: Listening on dbus.socket - D-Bus User Message Bus Socket. Dec 12 18:38:41.310098 systemd[1703]: Reached target sockets.target - Sockets. Dec 12 18:38:41.310239 systemd[1703]: Reached target basic.target - Basic System. Dec 12 18:38:41.310344 systemd[1703]: Reached target default.target - Main User Target. Dec 12 18:38:41.310426 systemd[1703]: Startup finished in 155ms. Dec 12 18:38:41.310744 systemd[1]: Started user@500.service - User Manager for UID 500. Dec 12 18:38:41.318472 systemd[1]: Started session-1.scope - Session 1 of User core. Dec 12 18:38:42.006488 systemd[1]: Started sshd@1-65.21.248.196:22-147.75.109.163:34830.service - OpenSSH per-connection server daemon (147.75.109.163:34830). Dec 12 18:38:42.992926 sshd[1714]: Accepted publickey for core from 147.75.109.163 port 34830 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:38:42.995061 sshd-session[1714]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:38:43.004795 systemd-logind[1520]: New session 2 of user core. Dec 12 18:38:43.011584 systemd[1]: Started session-2.scope - Session 2 of User core. Dec 12 18:38:43.670312 sshd[1717]: Connection closed by 147.75.109.163 port 34830 Dec 12 18:38:43.670891 sshd-session[1714]: pam_unix(sshd:session): session closed for user core Dec 12 18:38:43.674256 systemd[1]: sshd@1-65.21.248.196:22-147.75.109.163:34830.service: Deactivated successfully. Dec 12 18:38:43.676061 systemd[1]: session-2.scope: Deactivated successfully. Dec 12 18:38:43.677769 systemd-logind[1520]: Session 2 logged out. Waiting for processes to exit. Dec 12 18:38:43.678909 systemd-logind[1520]: Removed session 2. Dec 12 18:38:43.838843 systemd[1]: Started sshd@2-65.21.248.196:22-147.75.109.163:58132.service - OpenSSH per-connection server daemon (147.75.109.163:58132). Dec 12 18:38:44.825870 sshd[1723]: Accepted publickey for core from 147.75.109.163 port 58132 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:38:44.827856 sshd-session[1723]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:38:44.836975 systemd-logind[1520]: New session 3 of user core. Dec 12 18:38:44.839454 systemd[1]: Started session-3.scope - Session 3 of User core. Dec 12 18:38:45.500439 sshd[1726]: Connection closed by 147.75.109.163 port 58132 Dec 12 18:38:45.501334 sshd-session[1723]: pam_unix(sshd:session): session closed for user core Dec 12 18:38:45.506242 systemd[1]: sshd@2-65.21.248.196:22-147.75.109.163:58132.service: Deactivated successfully. Dec 12 18:38:45.508745 systemd[1]: session-3.scope: Deactivated successfully. Dec 12 18:38:45.511595 systemd-logind[1520]: Session 3 logged out. Waiting for processes to exit. Dec 12 18:38:45.513501 systemd-logind[1520]: Removed session 3. Dec 12 18:38:45.678122 systemd[1]: Started sshd@3-65.21.248.196:22-147.75.109.163:58148.service - OpenSSH per-connection server daemon (147.75.109.163:58148). Dec 12 18:38:46.685897 sshd[1732]: Accepted publickey for core from 147.75.109.163 port 58148 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:38:46.687408 sshd-session[1732]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:38:46.693420 systemd-logind[1520]: New session 4 of user core. Dec 12 18:38:46.705357 systemd[1]: Started session-4.scope - Session 4 of User core. Dec 12 18:38:47.203141 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Dec 12 18:38:47.207625 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:38:47.354153 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:38:47.360458 (kubelet)[1745]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 12 18:38:47.367218 sshd[1735]: Connection closed by 147.75.109.163 port 58148 Dec 12 18:38:47.367260 sshd-session[1732]: pam_unix(sshd:session): session closed for user core Dec 12 18:38:47.372407 systemd[1]: sshd@3-65.21.248.196:22-147.75.109.163:58148.service: Deactivated successfully. Dec 12 18:38:47.374131 systemd[1]: session-4.scope: Deactivated successfully. Dec 12 18:38:47.375603 systemd-logind[1520]: Session 4 logged out. Waiting for processes to exit. Dec 12 18:38:47.377862 systemd-logind[1520]: Removed session 4. Dec 12 18:38:47.400277 kubelet[1745]: E1212 18:38:47.400239 1745 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 12 18:38:47.403748 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 12 18:38:47.403885 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 12 18:38:47.404350 systemd[1]: kubelet.service: Consumed 159ms CPU time, 107.9M memory peak. Dec 12 18:38:47.542469 systemd[1]: Started sshd@4-65.21.248.196:22-147.75.109.163:58164.service - OpenSSH per-connection server daemon (147.75.109.163:58164). Dec 12 18:38:48.543846 sshd[1756]: Accepted publickey for core from 147.75.109.163 port 58164 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:38:48.545546 sshd-session[1756]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:38:48.550800 systemd-logind[1520]: New session 5 of user core. Dec 12 18:38:48.556372 systemd[1]: Started session-5.scope - Session 5 of User core. Dec 12 18:38:49.069533 sudo[1760]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Dec 12 18:38:49.069738 sudo[1760]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 12 18:38:49.083272 sudo[1760]: pam_unix(sudo:session): session closed for user root Dec 12 18:38:49.241596 sshd[1759]: Connection closed by 147.75.109.163 port 58164 Dec 12 18:38:49.242525 sshd-session[1756]: pam_unix(sshd:session): session closed for user core Dec 12 18:38:49.247462 systemd[1]: sshd@4-65.21.248.196:22-147.75.109.163:58164.service: Deactivated successfully. Dec 12 18:38:49.247818 systemd-logind[1520]: Session 5 logged out. Waiting for processes to exit. Dec 12 18:38:49.249522 systemd[1]: session-5.scope: Deactivated successfully. Dec 12 18:38:49.250985 systemd-logind[1520]: Removed session 5. Dec 12 18:38:49.413275 systemd[1]: Started sshd@5-65.21.248.196:22-147.75.109.163:58176.service - OpenSSH per-connection server daemon (147.75.109.163:58176). Dec 12 18:38:50.398909 sshd[1766]: Accepted publickey for core from 147.75.109.163 port 58176 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:38:50.400224 sshd-session[1766]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:38:50.404861 systemd-logind[1520]: New session 6 of user core. Dec 12 18:38:50.410367 systemd[1]: Started session-6.scope - Session 6 of User core. Dec 12 18:38:50.920732 sudo[1771]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Dec 12 18:38:50.921146 sudo[1771]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 12 18:38:50.927381 sudo[1771]: pam_unix(sudo:session): session closed for user root Dec 12 18:38:50.933081 sudo[1770]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Dec 12 18:38:50.933391 sudo[1770]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 12 18:38:50.947230 systemd[1]: Starting audit-rules.service - Load Audit Rules... Dec 12 18:38:50.996608 augenrules[1793]: No rules Dec 12 18:38:50.998030 systemd[1]: audit-rules.service: Deactivated successfully. Dec 12 18:38:50.998380 systemd[1]: Finished audit-rules.service - Load Audit Rules. Dec 12 18:38:50.999717 sudo[1770]: pam_unix(sudo:session): session closed for user root Dec 12 18:38:51.158434 sshd[1769]: Connection closed by 147.75.109.163 port 58176 Dec 12 18:38:51.158985 sshd-session[1766]: pam_unix(sshd:session): session closed for user core Dec 12 18:38:51.165304 systemd[1]: sshd@5-65.21.248.196:22-147.75.109.163:58176.service: Deactivated successfully. Dec 12 18:38:51.165408 systemd-logind[1520]: Session 6 logged out. Waiting for processes to exit. Dec 12 18:38:51.168742 systemd[1]: session-6.scope: Deactivated successfully. Dec 12 18:38:51.171343 systemd-logind[1520]: Removed session 6. Dec 12 18:38:51.336813 systemd[1]: Started sshd@6-65.21.248.196:22-147.75.109.163:58180.service - OpenSSH per-connection server daemon (147.75.109.163:58180). Dec 12 18:38:52.325479 sshd[1802]: Accepted publickey for core from 147.75.109.163 port 58180 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:38:52.326690 sshd-session[1802]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:38:52.331320 systemd-logind[1520]: New session 7 of user core. Dec 12 18:38:52.346358 systemd[1]: Started session-7.scope - Session 7 of User core. Dec 12 18:38:52.844872 sudo[1806]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 12 18:38:52.845214 sudo[1806]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 12 18:38:53.143142 systemd[1]: Starting docker.service - Docker Application Container Engine... Dec 12 18:38:53.157446 (dockerd)[1824]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Dec 12 18:38:53.361376 dockerd[1824]: time="2025-12-12T18:38:53.361145158Z" level=info msg="Starting up" Dec 12 18:38:53.362273 dockerd[1824]: time="2025-12-12T18:38:53.362256152Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Dec 12 18:38:53.373101 dockerd[1824]: time="2025-12-12T18:38:53.373019153Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Dec 12 18:38:53.387399 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2944607000-merged.mount: Deactivated successfully. Dec 12 18:38:53.416916 dockerd[1824]: time="2025-12-12T18:38:53.416828120Z" level=info msg="Loading containers: start." Dec 12 18:38:53.429220 kernel: Initializing XFRM netlink socket Dec 12 18:38:53.687629 systemd-networkd[1447]: docker0: Link UP Dec 12 18:38:53.693580 dockerd[1824]: time="2025-12-12T18:38:53.693518232Z" level=info msg="Loading containers: done." Dec 12 18:38:53.711841 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1605605912-merged.mount: Deactivated successfully. Dec 12 18:38:53.713163 dockerd[1824]: time="2025-12-12T18:38:53.713106531Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Dec 12 18:38:53.713356 dockerd[1824]: time="2025-12-12T18:38:53.713327797Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Dec 12 18:38:53.713447 dockerd[1824]: time="2025-12-12T18:38:53.713416062Z" level=info msg="Initializing buildkit" Dec 12 18:38:53.743466 dockerd[1824]: time="2025-12-12T18:38:53.743394004Z" level=info msg="Completed buildkit initialization" Dec 12 18:38:53.756728 dockerd[1824]: time="2025-12-12T18:38:53.756669878Z" level=info msg="Daemon has completed initialization" Dec 12 18:38:53.756934 dockerd[1824]: time="2025-12-12T18:38:53.756889680Z" level=info msg="API listen on /run/docker.sock" Dec 12 18:38:53.757013 systemd[1]: Started docker.service - Docker Application Container Engine. Dec 12 18:38:55.178099 containerd[1549]: time="2025-12-12T18:38:55.178048778Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.7\"" Dec 12 18:38:55.698567 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1743135259.mount: Deactivated successfully. Dec 12 18:38:56.714071 containerd[1549]: time="2025-12-12T18:38:56.714015341Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:56.715031 containerd[1549]: time="2025-12-12T18:38:56.714821884Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.7: active requests=0, bytes read=30114812" Dec 12 18:38:56.715879 containerd[1549]: time="2025-12-12T18:38:56.715847717Z" level=info msg="ImageCreate event name:\"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:56.717988 containerd[1549]: time="2025-12-12T18:38:56.717960138Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:9585226cb85d1dc0f0ef5f7a75f04e4bc91ddd82de249533bd293aa3cf958dab\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:56.718663 containerd[1549]: time="2025-12-12T18:38:56.718642718Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.7\" with image id \"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.7\", repo digest \"registry.k8s.io/kube-apiserver@sha256:9585226cb85d1dc0f0ef5f7a75f04e4bc91ddd82de249533bd293aa3cf958dab\", size \"30111311\" in 1.540538335s" Dec 12 18:38:56.718734 containerd[1549]: time="2025-12-12T18:38:56.718722378Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.7\" returns image reference \"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\"" Dec 12 18:38:56.719431 containerd[1549]: time="2025-12-12T18:38:56.719393225Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.7\"" Dec 12 18:38:57.452906 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Dec 12 18:38:57.455761 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:38:57.575298 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:38:57.591793 (kubelet)[2101]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 12 18:38:57.639960 kubelet[2101]: E1212 18:38:57.639925 2101 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 12 18:38:57.643729 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 12 18:38:57.643868 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 12 18:38:57.644575 systemd[1]: kubelet.service: Consumed 135ms CPU time, 108.5M memory peak. Dec 12 18:38:58.061212 containerd[1549]: time="2025-12-12T18:38:58.061135609Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:58.062113 containerd[1549]: time="2025-12-12T18:38:58.062086482Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.7: active requests=0, bytes read=26016803" Dec 12 18:38:58.063133 containerd[1549]: time="2025-12-12T18:38:58.062848451Z" level=info msg="ImageCreate event name:\"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:58.065599 containerd[1549]: time="2025-12-12T18:38:58.065570956Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f69d77ca0626b5a4b7b432c18de0952941181db7341c80eb89731f46d1d0c230\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:58.066450 containerd[1549]: time="2025-12-12T18:38:58.066431279Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.7\" with image id \"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.7\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f69d77ca0626b5a4b7b432c18de0952941181db7341c80eb89731f46d1d0c230\", size \"27673815\" in 1.347011915s" Dec 12 18:38:58.066520 containerd[1549]: time="2025-12-12T18:38:58.066508173Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.7\" returns image reference \"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\"" Dec 12 18:38:58.067394 containerd[1549]: time="2025-12-12T18:38:58.067335725Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.7\"" Dec 12 18:38:59.131010 containerd[1549]: time="2025-12-12T18:38:59.130946323Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:59.132462 containerd[1549]: time="2025-12-12T18:38:59.132421088Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.7: active requests=0, bytes read=20158124" Dec 12 18:38:59.133317 containerd[1549]: time="2025-12-12T18:38:59.133236447Z" level=info msg="ImageCreate event name:\"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:59.136237 containerd[1549]: time="2025-12-12T18:38:59.136146053Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:21bda321d8b4d48eb059fbc1593203d55d8b3bc7acd0584e04e55504796d78d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:38:59.137500 containerd[1549]: time="2025-12-12T18:38:59.137312350Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.7\" with image id \"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.7\", repo digest \"registry.k8s.io/kube-scheduler@sha256:21bda321d8b4d48eb059fbc1593203d55d8b3bc7acd0584e04e55504796d78d0\", size \"21815154\" in 1.069937232s" Dec 12 18:38:59.137500 containerd[1549]: time="2025-12-12T18:38:59.137349890Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.7\" returns image reference \"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\"" Dec 12 18:38:59.138294 containerd[1549]: time="2025-12-12T18:38:59.138251020Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.7\"" Dec 12 18:39:00.046150 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount811352611.mount: Deactivated successfully. Dec 12 18:39:00.337328 containerd[1549]: time="2025-12-12T18:39:00.337151150Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:00.338746 containerd[1549]: time="2025-12-12T18:39:00.338520157Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.7: active requests=0, bytes read=31930124" Dec 12 18:39:00.339727 containerd[1549]: time="2025-12-12T18:39:00.339688638Z" level=info msg="ImageCreate event name:\"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:00.342255 containerd[1549]: time="2025-12-12T18:39:00.342223702Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ec25702b19026e9c0d339bc1c3bd231435a59f28b5fccb21e1b1078a357380f5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:00.343062 containerd[1549]: time="2025-12-12T18:39:00.343018422Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.7\" with image id \"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\", repo tag \"registry.k8s.io/kube-proxy:v1.33.7\", repo digest \"registry.k8s.io/kube-proxy@sha256:ec25702b19026e9c0d339bc1c3bd231435a59f28b5fccb21e1b1078a357380f5\", size \"31929115\" in 1.204731805s" Dec 12 18:39:00.343154 containerd[1549]: time="2025-12-12T18:39:00.343134108Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.7\" returns image reference \"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\"" Dec 12 18:39:00.343971 containerd[1549]: time="2025-12-12T18:39:00.343933718Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Dec 12 18:39:00.818855 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount125686400.mount: Deactivated successfully. Dec 12 18:39:01.582540 containerd[1549]: time="2025-12-12T18:39:01.582499605Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:01.583538 containerd[1549]: time="2025-12-12T18:39:01.583317569Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942332" Dec 12 18:39:01.584228 containerd[1549]: time="2025-12-12T18:39:01.584175318Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:01.586380 containerd[1549]: time="2025-12-12T18:39:01.586359843Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:01.587009 containerd[1549]: time="2025-12-12T18:39:01.586981589Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.243016153s" Dec 12 18:39:01.587054 containerd[1549]: time="2025-12-12T18:39:01.587009802Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Dec 12 18:39:01.587472 containerd[1549]: time="2025-12-12T18:39:01.587430361Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Dec 12 18:39:02.015674 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4198593240.mount: Deactivated successfully. Dec 12 18:39:02.020477 containerd[1549]: time="2025-12-12T18:39:02.020407457Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 12 18:39:02.021354 containerd[1549]: time="2025-12-12T18:39:02.021236682Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321160" Dec 12 18:39:02.022231 containerd[1549]: time="2025-12-12T18:39:02.022162938Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 12 18:39:02.025004 containerd[1549]: time="2025-12-12T18:39:02.024304645Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 12 18:39:02.025004 containerd[1549]: time="2025-12-12T18:39:02.024887367Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 437.426058ms" Dec 12 18:39:02.025004 containerd[1549]: time="2025-12-12T18:39:02.024914217Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Dec 12 18:39:02.025660 containerd[1549]: time="2025-12-12T18:39:02.025616244Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Dec 12 18:39:02.511843 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1316105296.mount: Deactivated successfully. Dec 12 18:39:04.015846 containerd[1549]: time="2025-12-12T18:39:04.015789722Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:04.016993 containerd[1549]: time="2025-12-12T18:39:04.016953023Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58926291" Dec 12 18:39:04.017940 containerd[1549]: time="2025-12-12T18:39:04.017899828Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:04.019979 containerd[1549]: time="2025-12-12T18:39:04.019932170Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:04.021162 containerd[1549]: time="2025-12-12T18:39:04.020841444Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 1.995197308s" Dec 12 18:39:04.021162 containerd[1549]: time="2025-12-12T18:39:04.020868505Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Dec 12 18:39:07.141526 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:39:07.141705 systemd[1]: kubelet.service: Consumed 135ms CPU time, 108.5M memory peak. Dec 12 18:39:07.144235 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:39:07.179056 systemd[1]: Reload requested from client PID 2263 ('systemctl') (unit session-7.scope)... Dec 12 18:39:07.179074 systemd[1]: Reloading... Dec 12 18:39:07.269237 zram_generator::config[2307]: No configuration found. Dec 12 18:39:07.435086 systemd[1]: Reloading finished in 255 ms. Dec 12 18:39:07.483838 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Dec 12 18:39:07.483918 systemd[1]: kubelet.service: Failed with result 'signal'. Dec 12 18:39:07.484117 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:39:07.484174 systemd[1]: kubelet.service: Consumed 71ms CPU time, 98.4M memory peak. Dec 12 18:39:07.486365 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:39:07.580954 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:39:07.586493 (kubelet)[2362]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 12 18:39:07.631925 kubelet[2362]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 18:39:07.632297 kubelet[2362]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Dec 12 18:39:07.632348 kubelet[2362]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 18:39:07.632461 kubelet[2362]: I1212 18:39:07.632434 2362 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 12 18:39:08.031107 kubelet[2362]: I1212 18:39:08.031075 2362 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Dec 12 18:39:08.031293 kubelet[2362]: I1212 18:39:08.031280 2362 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 12 18:39:08.031936 kubelet[2362]: I1212 18:39:08.031921 2362 server.go:956] "Client rotation is on, will bootstrap in background" Dec 12 18:39:08.077240 kubelet[2362]: I1212 18:39:08.076799 2362 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 12 18:39:08.081425 kubelet[2362]: E1212 18:39:08.081368 2362 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://65.21.248.196:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 65.21.248.196:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Dec 12 18:39:08.094751 kubelet[2362]: I1212 18:39:08.094717 2362 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 12 18:39:08.101850 kubelet[2362]: I1212 18:39:08.101815 2362 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 12 18:39:08.105784 kubelet[2362]: I1212 18:39:08.105725 2362 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 12 18:39:08.109203 kubelet[2362]: I1212 18:39:08.105764 2362 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459-2-2-7-adcd881523","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 12 18:39:08.109203 kubelet[2362]: I1212 18:39:08.109202 2362 topology_manager.go:138] "Creating topology manager with none policy" Dec 12 18:39:08.109344 kubelet[2362]: I1212 18:39:08.109214 2362 container_manager_linux.go:303] "Creating device plugin manager" Dec 12 18:39:08.110166 kubelet[2362]: I1212 18:39:08.110138 2362 state_mem.go:36] "Initialized new in-memory state store" Dec 12 18:39:08.116566 kubelet[2362]: I1212 18:39:08.115297 2362 kubelet.go:480] "Attempting to sync node with API server" Dec 12 18:39:08.116566 kubelet[2362]: I1212 18:39:08.115318 2362 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 12 18:39:08.116566 kubelet[2362]: I1212 18:39:08.115348 2362 kubelet.go:386] "Adding apiserver pod source" Dec 12 18:39:08.117684 kubelet[2362]: I1212 18:39:08.117642 2362 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 12 18:39:08.123042 kubelet[2362]: E1212 18:39:08.123020 2362 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://65.21.248.196:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4459-2-2-7-adcd881523&limit=500&resourceVersion=0\": dial tcp 65.21.248.196:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 12 18:39:08.123844 kubelet[2362]: I1212 18:39:08.123238 2362 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Dec 12 18:39:08.123844 kubelet[2362]: I1212 18:39:08.123775 2362 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Dec 12 18:39:08.124945 kubelet[2362]: W1212 18:39:08.124923 2362 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 12 18:39:08.135147 kubelet[2362]: I1212 18:39:08.135116 2362 watchdog_linux.go:99] "Systemd watchdog is not enabled" Dec 12 18:39:08.135250 kubelet[2362]: I1212 18:39:08.135213 2362 server.go:1289] "Started kubelet" Dec 12 18:39:08.136921 kubelet[2362]: E1212 18:39:08.136833 2362 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://65.21.248.196:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 65.21.248.196:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 12 18:39:08.138034 kubelet[2362]: I1212 18:39:08.137400 2362 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Dec 12 18:39:08.142150 kubelet[2362]: I1212 18:39:08.141923 2362 server.go:317] "Adding debug handlers to kubelet server" Dec 12 18:39:08.143221 kubelet[2362]: I1212 18:39:08.142791 2362 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 12 18:39:08.143221 kubelet[2362]: I1212 18:39:08.143089 2362 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 12 18:39:08.145204 kubelet[2362]: I1212 18:39:08.144378 2362 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 12 18:39:08.148749 kubelet[2362]: E1212 18:39:08.145582 2362 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://65.21.248.196:6443/api/v1/namespaces/default/events\": dial tcp 65.21.248.196:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4459-2-2-7-adcd881523.18808bc8b88bfa43 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4459-2-2-7-adcd881523,UID:ci-4459-2-2-7-adcd881523,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4459-2-2-7-adcd881523,},FirstTimestamp:2025-12-12 18:39:08.135135811 +0000 UTC m=+0.544351563,LastTimestamp:2025-12-12 18:39:08.135135811 +0000 UTC m=+0.544351563,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459-2-2-7-adcd881523,}" Dec 12 18:39:08.148749 kubelet[2362]: I1212 18:39:08.148650 2362 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 12 18:39:08.152468 kubelet[2362]: E1212 18:39:08.151457 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4459-2-2-7-adcd881523\" not found" Dec 12 18:39:08.152468 kubelet[2362]: I1212 18:39:08.151499 2362 volume_manager.go:297] "Starting Kubelet Volume Manager" Dec 12 18:39:08.152468 kubelet[2362]: I1212 18:39:08.151694 2362 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Dec 12 18:39:08.152468 kubelet[2362]: I1212 18:39:08.151741 2362 reconciler.go:26] "Reconciler: start to sync state" Dec 12 18:39:08.152468 kubelet[2362]: E1212 18:39:08.152088 2362 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://65.21.248.196:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 65.21.248.196:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 12 18:39:08.154248 kubelet[2362]: E1212 18:39:08.154213 2362 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://65.21.248.196:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-2-2-7-adcd881523?timeout=10s\": dial tcp 65.21.248.196:6443: connect: connection refused" interval="200ms" Dec 12 18:39:08.156412 kubelet[2362]: I1212 18:39:08.156382 2362 factory.go:223] Registration of the systemd container factory successfully Dec 12 18:39:08.156474 kubelet[2362]: I1212 18:39:08.156452 2362 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 12 18:39:08.159818 kubelet[2362]: I1212 18:39:08.159806 2362 factory.go:223] Registration of the containerd container factory successfully Dec 12 18:39:08.164692 kubelet[2362]: E1212 18:39:08.164664 2362 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 12 18:39:08.179367 kubelet[2362]: I1212 18:39:08.177778 2362 cpu_manager.go:221] "Starting CPU manager" policy="none" Dec 12 18:39:08.179367 kubelet[2362]: I1212 18:39:08.177791 2362 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Dec 12 18:39:08.179367 kubelet[2362]: I1212 18:39:08.177805 2362 state_mem.go:36] "Initialized new in-memory state store" Dec 12 18:39:08.179824 kubelet[2362]: I1212 18:39:08.179753 2362 policy_none.go:49] "None policy: Start" Dec 12 18:39:08.179824 kubelet[2362]: I1212 18:39:08.179797 2362 memory_manager.go:186] "Starting memorymanager" policy="None" Dec 12 18:39:08.179824 kubelet[2362]: I1212 18:39:08.179813 2362 state_mem.go:35] "Initializing new in-memory state store" Dec 12 18:39:08.182045 kubelet[2362]: I1212 18:39:08.181134 2362 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Dec 12 18:39:08.184526 kubelet[2362]: I1212 18:39:08.184507 2362 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Dec 12 18:39:08.184616 kubelet[2362]: I1212 18:39:08.184599 2362 status_manager.go:230] "Starting to sync pod status with apiserver" Dec 12 18:39:08.184665 kubelet[2362]: I1212 18:39:08.184635 2362 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Dec 12 18:39:08.184665 kubelet[2362]: I1212 18:39:08.184644 2362 kubelet.go:2436] "Starting kubelet main sync loop" Dec 12 18:39:08.184729 kubelet[2362]: E1212 18:39:08.184701 2362 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 12 18:39:08.187944 kubelet[2362]: E1212 18:39:08.187895 2362 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://65.21.248.196:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 65.21.248.196:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 12 18:39:08.193078 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Dec 12 18:39:08.198965 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Dec 12 18:39:08.202131 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Dec 12 18:39:08.211858 kubelet[2362]: E1212 18:39:08.211808 2362 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Dec 12 18:39:08.212085 kubelet[2362]: I1212 18:39:08.212006 2362 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 12 18:39:08.212085 kubelet[2362]: I1212 18:39:08.212039 2362 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 12 18:39:08.212538 kubelet[2362]: I1212 18:39:08.212434 2362 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 12 18:39:08.213981 kubelet[2362]: E1212 18:39:08.213965 2362 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Dec 12 18:39:08.214120 kubelet[2362]: E1212 18:39:08.214093 2362 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4459-2-2-7-adcd881523\" not found" Dec 12 18:39:08.272318 kubelet[2362]: E1212 18:39:08.272217 2362 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://65.21.248.196:6443/api/v1/namespaces/default/events\": dial tcp 65.21.248.196:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4459-2-2-7-adcd881523.18808bc8b88bfa43 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4459-2-2-7-adcd881523,UID:ci-4459-2-2-7-adcd881523,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4459-2-2-7-adcd881523,},FirstTimestamp:2025-12-12 18:39:08.135135811 +0000 UTC m=+0.544351563,LastTimestamp:2025-12-12 18:39:08.135135811 +0000 UTC m=+0.544351563,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459-2-2-7-adcd881523,}" Dec 12 18:39:08.298376 systemd[1]: Created slice kubepods-burstable-podc5a27f6c1bcf716779a9e59fcfc8c831.slice - libcontainer container kubepods-burstable-podc5a27f6c1bcf716779a9e59fcfc8c831.slice. Dec 12 18:39:08.306828 kubelet[2362]: E1212 18:39:08.306781 2362 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.310524 systemd[1]: Created slice kubepods-burstable-poddc0e16563336e6225c3aecdefa7940f4.slice - libcontainer container kubepods-burstable-poddc0e16563336e6225c3aecdefa7940f4.slice. Dec 12 18:39:08.314893 kubelet[2362]: I1212 18:39:08.314867 2362 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.315218 kubelet[2362]: E1212 18:39:08.315149 2362 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://65.21.248.196:6443/api/v1/nodes\": dial tcp 65.21.248.196:6443: connect: connection refused" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.319288 kubelet[2362]: E1212 18:39:08.319070 2362 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.322256 systemd[1]: Created slice kubepods-burstable-podecd9d4e6c082fd45c85023240c932785.slice - libcontainer container kubepods-burstable-podecd9d4e6c082fd45c85023240c932785.slice. Dec 12 18:39:08.324016 kubelet[2362]: E1212 18:39:08.323999 2362 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353309 kubelet[2362]: I1212 18:39:08.353279 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-kubeconfig\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353391 kubelet[2362]: I1212 18:39:08.353330 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ecd9d4e6c082fd45c85023240c932785-kubeconfig\") pod \"kube-scheduler-ci-4459-2-2-7-adcd881523\" (UID: \"ecd9d4e6c082fd45c85023240c932785\") " pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353391 kubelet[2362]: I1212 18:39:08.353372 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c5a27f6c1bcf716779a9e59fcfc8c831-ca-certs\") pod \"kube-apiserver-ci-4459-2-2-7-adcd881523\" (UID: \"c5a27f6c1bcf716779a9e59fcfc8c831\") " pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353446 kubelet[2362]: I1212 18:39:08.353396 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c5a27f6c1bcf716779a9e59fcfc8c831-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459-2-2-7-adcd881523\" (UID: \"c5a27f6c1bcf716779a9e59fcfc8c831\") " pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353446 kubelet[2362]: I1212 18:39:08.353417 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-flexvolume-dir\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353446 kubelet[2362]: I1212 18:39:08.353432 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-k8s-certs\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353506 kubelet[2362]: I1212 18:39:08.353454 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353506 kubelet[2362]: I1212 18:39:08.353468 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c5a27f6c1bcf716779a9e59fcfc8c831-k8s-certs\") pod \"kube-apiserver-ci-4459-2-2-7-adcd881523\" (UID: \"c5a27f6c1bcf716779a9e59fcfc8c831\") " pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.353506 kubelet[2362]: I1212 18:39:08.353481 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-ca-certs\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.354654 kubelet[2362]: E1212 18:39:08.354622 2362 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://65.21.248.196:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-2-2-7-adcd881523?timeout=10s\": dial tcp 65.21.248.196:6443: connect: connection refused" interval="400ms" Dec 12 18:39:08.519079 kubelet[2362]: I1212 18:39:08.519033 2362 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.519393 kubelet[2362]: E1212 18:39:08.519356 2362 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://65.21.248.196:6443/api/v1/nodes\": dial tcp 65.21.248.196:6443: connect: connection refused" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.609446 containerd[1549]: time="2025-12-12T18:39:08.609322606Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459-2-2-7-adcd881523,Uid:c5a27f6c1bcf716779a9e59fcfc8c831,Namespace:kube-system,Attempt:0,}" Dec 12 18:39:08.620823 containerd[1549]: time="2025-12-12T18:39:08.619801274Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459-2-2-7-adcd881523,Uid:dc0e16563336e6225c3aecdefa7940f4,Namespace:kube-system,Attempt:0,}" Dec 12 18:39:08.638655 containerd[1549]: time="2025-12-12T18:39:08.638588852Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459-2-2-7-adcd881523,Uid:ecd9d4e6c082fd45c85023240c932785,Namespace:kube-system,Attempt:0,}" Dec 12 18:39:08.696679 containerd[1549]: time="2025-12-12T18:39:08.696608109Z" level=info msg="connecting to shim 028242f3d5565cee9a64dea179a7836e5fd3d679545ad9b9133b978fb488c654" address="unix:///run/containerd/s/8bfa28b7d89a64e05d1b64ecc381118790e6658b81822536587c027fd3177bb9" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:08.697083 containerd[1549]: time="2025-12-12T18:39:08.697018695Z" level=info msg="connecting to shim 6fbff13c9b911b11a72812bd7cf181a943aef534a19f8dd9e42d17dc7a73d2cb" address="unix:///run/containerd/s/d70c1b4eef1f55a70d192310b96baadd1634a85fa5272105e8ec1ecbe291a2d7" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:08.699852 containerd[1549]: time="2025-12-12T18:39:08.699828702Z" level=info msg="connecting to shim f4e74119beee54d61f4c88ee055485b6bf612c027040f409e463a50b11d392d5" address="unix:///run/containerd/s/394de8a79724ef5e1ca44de206acbc1eb3db5d6b6632f889be6185e8adb53247" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:08.757931 kubelet[2362]: E1212 18:39:08.756696 2362 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://65.21.248.196:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-2-2-7-adcd881523?timeout=10s\": dial tcp 65.21.248.196:6443: connect: connection refused" interval="800ms" Dec 12 18:39:08.773301 systemd[1]: Started cri-containerd-028242f3d5565cee9a64dea179a7836e5fd3d679545ad9b9133b978fb488c654.scope - libcontainer container 028242f3d5565cee9a64dea179a7836e5fd3d679545ad9b9133b978fb488c654. Dec 12 18:39:08.774047 systemd[1]: Started cri-containerd-6fbff13c9b911b11a72812bd7cf181a943aef534a19f8dd9e42d17dc7a73d2cb.scope - libcontainer container 6fbff13c9b911b11a72812bd7cf181a943aef534a19f8dd9e42d17dc7a73d2cb. Dec 12 18:39:08.774874 systemd[1]: Started cri-containerd-f4e74119beee54d61f4c88ee055485b6bf612c027040f409e463a50b11d392d5.scope - libcontainer container f4e74119beee54d61f4c88ee055485b6bf612c027040f409e463a50b11d392d5. Dec 12 18:39:08.834595 containerd[1549]: time="2025-12-12T18:39:08.834559026Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459-2-2-7-adcd881523,Uid:c5a27f6c1bcf716779a9e59fcfc8c831,Namespace:kube-system,Attempt:0,} returns sandbox id \"f4e74119beee54d61f4c88ee055485b6bf612c027040f409e463a50b11d392d5\"" Dec 12 18:39:08.841565 containerd[1549]: time="2025-12-12T18:39:08.841546391Z" level=info msg="CreateContainer within sandbox \"f4e74119beee54d61f4c88ee055485b6bf612c027040f409e463a50b11d392d5\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Dec 12 18:39:08.843552 containerd[1549]: time="2025-12-12T18:39:08.843532351Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459-2-2-7-adcd881523,Uid:dc0e16563336e6225c3aecdefa7940f4,Namespace:kube-system,Attempt:0,} returns sandbox id \"028242f3d5565cee9a64dea179a7836e5fd3d679545ad9b9133b978fb488c654\"" Dec 12 18:39:08.846812 containerd[1549]: time="2025-12-12T18:39:08.846795964Z" level=info msg="CreateContainer within sandbox \"028242f3d5565cee9a64dea179a7836e5fd3d679545ad9b9133b978fb488c654\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Dec 12 18:39:08.854434 containerd[1549]: time="2025-12-12T18:39:08.854417037Z" level=info msg="Container 0b7393ab3289458a5765af3ba6837d19dfaedb067b82c5ddb02c291a1e460bb9: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:08.854690 containerd[1549]: time="2025-12-12T18:39:08.854671185Z" level=info msg="Container 95982280df3e30d0b166a7621dc3eaac25e93c5e8a1b3f7355f6c35f41573af6: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:08.866602 containerd[1549]: time="2025-12-12T18:39:08.866509038Z" level=info msg="CreateContainer within sandbox \"f4e74119beee54d61f4c88ee055485b6bf612c027040f409e463a50b11d392d5\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"95982280df3e30d0b166a7621dc3eaac25e93c5e8a1b3f7355f6c35f41573af6\"" Dec 12 18:39:08.868347 containerd[1549]: time="2025-12-12T18:39:08.868331735Z" level=info msg="StartContainer for \"95982280df3e30d0b166a7621dc3eaac25e93c5e8a1b3f7355f6c35f41573af6\"" Dec 12 18:39:08.870038 containerd[1549]: time="2025-12-12T18:39:08.869964905Z" level=info msg="CreateContainer within sandbox \"028242f3d5565cee9a64dea179a7836e5fd3d679545ad9b9133b978fb488c654\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"0b7393ab3289458a5765af3ba6837d19dfaedb067b82c5ddb02c291a1e460bb9\"" Dec 12 18:39:08.870577 containerd[1549]: time="2025-12-12T18:39:08.870562587Z" level=info msg="StartContainer for \"0b7393ab3289458a5765af3ba6837d19dfaedb067b82c5ddb02c291a1e460bb9\"" Dec 12 18:39:08.870879 containerd[1549]: time="2025-12-12T18:39:08.870839276Z" level=info msg="connecting to shim 95982280df3e30d0b166a7621dc3eaac25e93c5e8a1b3f7355f6c35f41573af6" address="unix:///run/containerd/s/394de8a79724ef5e1ca44de206acbc1eb3db5d6b6632f889be6185e8adb53247" protocol=ttrpc version=3 Dec 12 18:39:08.871753 containerd[1549]: time="2025-12-12T18:39:08.871728736Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459-2-2-7-adcd881523,Uid:ecd9d4e6c082fd45c85023240c932785,Namespace:kube-system,Attempt:0,} returns sandbox id \"6fbff13c9b911b11a72812bd7cf181a943aef534a19f8dd9e42d17dc7a73d2cb\"" Dec 12 18:39:08.872089 containerd[1549]: time="2025-12-12T18:39:08.871899019Z" level=info msg="connecting to shim 0b7393ab3289458a5765af3ba6837d19dfaedb067b82c5ddb02c291a1e460bb9" address="unix:///run/containerd/s/8bfa28b7d89a64e05d1b64ecc381118790e6658b81822536587c027fd3177bb9" protocol=ttrpc version=3 Dec 12 18:39:08.877485 containerd[1549]: time="2025-12-12T18:39:08.877465155Z" level=info msg="CreateContainer within sandbox \"6fbff13c9b911b11a72812bd7cf181a943aef534a19f8dd9e42d17dc7a73d2cb\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Dec 12 18:39:08.892313 systemd[1]: Started cri-containerd-0b7393ab3289458a5765af3ba6837d19dfaedb067b82c5ddb02c291a1e460bb9.scope - libcontainer container 0b7393ab3289458a5765af3ba6837d19dfaedb067b82c5ddb02c291a1e460bb9. Dec 12 18:39:08.893598 systemd[1]: Started cri-containerd-95982280df3e30d0b166a7621dc3eaac25e93c5e8a1b3f7355f6c35f41573af6.scope - libcontainer container 95982280df3e30d0b166a7621dc3eaac25e93c5e8a1b3f7355f6c35f41573af6. Dec 12 18:39:08.894962 containerd[1549]: time="2025-12-12T18:39:08.894890223Z" level=info msg="Container 0c9b8365decd9a8c79a4e77a88429d2af01be69508fd614dacfb50322c0064d5: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:08.904589 containerd[1549]: time="2025-12-12T18:39:08.904558617Z" level=info msg="CreateContainer within sandbox \"6fbff13c9b911b11a72812bd7cf181a943aef534a19f8dd9e42d17dc7a73d2cb\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"0c9b8365decd9a8c79a4e77a88429d2af01be69508fd614dacfb50322c0064d5\"" Dec 12 18:39:08.904957 containerd[1549]: time="2025-12-12T18:39:08.904932116Z" level=info msg="StartContainer for \"0c9b8365decd9a8c79a4e77a88429d2af01be69508fd614dacfb50322c0064d5\"" Dec 12 18:39:08.908796 containerd[1549]: time="2025-12-12T18:39:08.908771671Z" level=info msg="connecting to shim 0c9b8365decd9a8c79a4e77a88429d2af01be69508fd614dacfb50322c0064d5" address="unix:///run/containerd/s/d70c1b4eef1f55a70d192310b96baadd1634a85fa5272105e8ec1ecbe291a2d7" protocol=ttrpc version=3 Dec 12 18:39:08.924541 kubelet[2362]: I1212 18:39:08.924515 2362 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.924786 kubelet[2362]: E1212 18:39:08.924769 2362 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://65.21.248.196:6443/api/v1/nodes\": dial tcp 65.21.248.196:6443: connect: connection refused" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:08.927401 systemd[1]: Started cri-containerd-0c9b8365decd9a8c79a4e77a88429d2af01be69508fd614dacfb50322c0064d5.scope - libcontainer container 0c9b8365decd9a8c79a4e77a88429d2af01be69508fd614dacfb50322c0064d5. Dec 12 18:39:08.955987 containerd[1549]: time="2025-12-12T18:39:08.955936577Z" level=info msg="StartContainer for \"95982280df3e30d0b166a7621dc3eaac25e93c5e8a1b3f7355f6c35f41573af6\" returns successfully" Dec 12 18:39:08.977891 containerd[1549]: time="2025-12-12T18:39:08.977849764Z" level=info msg="StartContainer for \"0b7393ab3289458a5765af3ba6837d19dfaedb067b82c5ddb02c291a1e460bb9\" returns successfully" Dec 12 18:39:09.002870 containerd[1549]: time="2025-12-12T18:39:09.002838221Z" level=info msg="StartContainer for \"0c9b8365decd9a8c79a4e77a88429d2af01be69508fd614dacfb50322c0064d5\" returns successfully" Dec 12 18:39:09.045202 kubelet[2362]: E1212 18:39:09.044523 2362 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://65.21.248.196:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 65.21.248.196:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 12 18:39:09.201450 kubelet[2362]: E1212 18:39:09.201235 2362 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:09.207795 kubelet[2362]: E1212 18:39:09.207771 2362 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:09.208714 kubelet[2362]: E1212 18:39:09.208693 2362 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:09.728286 kubelet[2362]: I1212 18:39:09.728254 2362 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.212231 kubelet[2362]: E1212 18:39:10.212210 2362 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.213201 kubelet[2362]: E1212 18:39:10.212535 2362 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.349385 kubelet[2362]: E1212 18:39:10.349342 2362 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4459-2-2-7-adcd881523\" not found" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.425384 kubelet[2362]: I1212 18:39:10.425354 2362 kubelet_node_status.go:78] "Successfully registered node" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.425384 kubelet[2362]: E1212 18:39:10.425384 2362 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-4459-2-2-7-adcd881523\": node \"ci-4459-2-2-7-adcd881523\" not found" Dec 12 18:39:10.452267 kubelet[2362]: I1212 18:39:10.452237 2362 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.459928 kubelet[2362]: E1212 18:39:10.459889 2362 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459-2-2-7-adcd881523\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.459928 kubelet[2362]: I1212 18:39:10.459911 2362 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.463545 kubelet[2362]: E1212 18:39:10.462854 2362 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.463545 kubelet[2362]: I1212 18:39:10.462873 2362 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:10.464155 kubelet[2362]: E1212 18:39:10.464130 2362 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459-2-2-7-adcd881523\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:11.132026 kubelet[2362]: I1212 18:39:11.131974 2362 apiserver.go:52] "Watching apiserver" Dec 12 18:39:11.152426 kubelet[2362]: I1212 18:39:11.152379 2362 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Dec 12 18:39:11.212778 kubelet[2362]: I1212 18:39:11.212722 2362 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:11.214679 kubelet[2362]: E1212 18:39:11.214639 2362 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459-2-2-7-adcd881523\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:11.579969 kubelet[2362]: I1212 18:39:11.579937 2362 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:12.540098 systemd[1]: Reload requested from client PID 2642 ('systemctl') (unit session-7.scope)... Dec 12 18:39:12.540118 systemd[1]: Reloading... Dec 12 18:39:12.613237 kubelet[2362]: I1212 18:39:12.611852 2362 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:12.626210 zram_generator::config[2682]: No configuration found. Dec 12 18:39:12.816399 systemd[1]: Reloading finished in 275 ms. Dec 12 18:39:12.842860 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:39:12.843637 kubelet[2362]: I1212 18:39:12.842969 2362 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 12 18:39:12.858491 systemd[1]: kubelet.service: Deactivated successfully. Dec 12 18:39:12.858806 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:39:12.858976 systemd[1]: kubelet.service: Consumed 839ms CPU time, 128.2M memory peak. Dec 12 18:39:12.860976 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:39:13.007482 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:39:13.014676 (kubelet)[2737]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 12 18:39:13.058273 kubelet[2737]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 18:39:13.058273 kubelet[2737]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Dec 12 18:39:13.058273 kubelet[2737]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 18:39:13.058608 kubelet[2737]: I1212 18:39:13.058303 2737 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 12 18:39:13.067126 kubelet[2737]: I1212 18:39:13.066908 2737 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Dec 12 18:39:13.067126 kubelet[2737]: I1212 18:39:13.066929 2737 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 12 18:39:13.068970 kubelet[2737]: I1212 18:39:13.067489 2737 server.go:956] "Client rotation is on, will bootstrap in background" Dec 12 18:39:13.069669 kubelet[2737]: I1212 18:39:13.069651 2737 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Dec 12 18:39:13.072210 kubelet[2737]: I1212 18:39:13.072170 2737 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 12 18:39:13.078201 kubelet[2737]: I1212 18:39:13.078159 2737 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 12 18:39:13.082747 kubelet[2737]: I1212 18:39:13.082733 2737 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 12 18:39:13.083022 kubelet[2737]: I1212 18:39:13.083000 2737 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 12 18:39:13.083246 kubelet[2737]: I1212 18:39:13.083090 2737 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459-2-2-7-adcd881523","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 12 18:39:13.083372 kubelet[2737]: I1212 18:39:13.083360 2737 topology_manager.go:138] "Creating topology manager with none policy" Dec 12 18:39:13.083425 kubelet[2737]: I1212 18:39:13.083418 2737 container_manager_linux.go:303] "Creating device plugin manager" Dec 12 18:39:13.083508 kubelet[2737]: I1212 18:39:13.083499 2737 state_mem.go:36] "Initialized new in-memory state store" Dec 12 18:39:13.083685 kubelet[2737]: I1212 18:39:13.083674 2737 kubelet.go:480] "Attempting to sync node with API server" Dec 12 18:39:13.083746 kubelet[2737]: I1212 18:39:13.083738 2737 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 12 18:39:13.083803 kubelet[2737]: I1212 18:39:13.083796 2737 kubelet.go:386] "Adding apiserver pod source" Dec 12 18:39:13.083860 kubelet[2737]: I1212 18:39:13.083852 2737 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 12 18:39:13.086484 kubelet[2737]: I1212 18:39:13.086396 2737 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Dec 12 18:39:13.086818 kubelet[2737]: I1212 18:39:13.086806 2737 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Dec 12 18:39:13.091858 kubelet[2737]: I1212 18:39:13.091688 2737 watchdog_linux.go:99] "Systemd watchdog is not enabled" Dec 12 18:39:13.091858 kubelet[2737]: I1212 18:39:13.091717 2737 server.go:1289] "Started kubelet" Dec 12 18:39:13.093710 kubelet[2737]: I1212 18:39:13.093694 2737 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 12 18:39:13.104263 kubelet[2737]: I1212 18:39:13.104239 2737 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Dec 12 18:39:13.105651 kubelet[2737]: I1212 18:39:13.105639 2737 server.go:317] "Adding debug handlers to kubelet server" Dec 12 18:39:13.111368 kubelet[2737]: I1212 18:39:13.111323 2737 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 12 18:39:13.111524 kubelet[2737]: I1212 18:39:13.111491 2737 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 12 18:39:13.111708 kubelet[2737]: I1212 18:39:13.111684 2737 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 12 18:39:13.112990 kubelet[2737]: I1212 18:39:13.112972 2737 volume_manager.go:297] "Starting Kubelet Volume Manager" Dec 12 18:39:13.113858 kubelet[2737]: I1212 18:39:13.113840 2737 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Dec 12 18:39:13.113935 kubelet[2737]: I1212 18:39:13.113923 2737 reconciler.go:26] "Reconciler: start to sync state" Dec 12 18:39:13.116119 kubelet[2737]: I1212 18:39:13.116091 2737 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Dec 12 18:39:13.117371 kubelet[2737]: I1212 18:39:13.117290 2737 factory.go:223] Registration of the systemd container factory successfully Dec 12 18:39:13.117371 kubelet[2737]: I1212 18:39:13.117314 2737 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Dec 12 18:39:13.117371 kubelet[2737]: I1212 18:39:13.117330 2737 status_manager.go:230] "Starting to sync pod status with apiserver" Dec 12 18:39:13.117371 kubelet[2737]: I1212 18:39:13.117344 2737 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Dec 12 18:39:13.117371 kubelet[2737]: I1212 18:39:13.117350 2737 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 12 18:39:13.117543 kubelet[2737]: I1212 18:39:13.117517 2737 kubelet.go:2436] "Starting kubelet main sync loop" Dec 12 18:39:13.117704 kubelet[2737]: E1212 18:39:13.117688 2737 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 12 18:39:13.119340 kubelet[2737]: I1212 18:39:13.118832 2737 factory.go:223] Registration of the containerd container factory successfully Dec 12 18:39:13.162504 kubelet[2737]: I1212 18:39:13.162476 2737 cpu_manager.go:221] "Starting CPU manager" policy="none" Dec 12 18:39:13.162504 kubelet[2737]: I1212 18:39:13.162492 2737 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Dec 12 18:39:13.162504 kubelet[2737]: I1212 18:39:13.162506 2737 state_mem.go:36] "Initialized new in-memory state store" Dec 12 18:39:13.162648 kubelet[2737]: I1212 18:39:13.162595 2737 state_mem.go:88] "Updated default CPUSet" cpuSet="" Dec 12 18:39:13.162648 kubelet[2737]: I1212 18:39:13.162604 2737 state_mem.go:96] "Updated CPUSet assignments" assignments={} Dec 12 18:39:13.162648 kubelet[2737]: I1212 18:39:13.162617 2737 policy_none.go:49] "None policy: Start" Dec 12 18:39:13.162648 kubelet[2737]: I1212 18:39:13.162624 2737 memory_manager.go:186] "Starting memorymanager" policy="None" Dec 12 18:39:13.162648 kubelet[2737]: I1212 18:39:13.162631 2737 state_mem.go:35] "Initializing new in-memory state store" Dec 12 18:39:13.162726 kubelet[2737]: I1212 18:39:13.162692 2737 state_mem.go:75] "Updated machine memory state" Dec 12 18:39:13.166034 kubelet[2737]: E1212 18:39:13.166008 2737 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Dec 12 18:39:13.166153 kubelet[2737]: I1212 18:39:13.166135 2737 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 12 18:39:13.166820 kubelet[2737]: I1212 18:39:13.166148 2737 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 12 18:39:13.167114 kubelet[2737]: I1212 18:39:13.167095 2737 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 12 18:39:13.172274 kubelet[2737]: E1212 18:39:13.172171 2737 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Dec 12 18:39:13.218646 kubelet[2737]: I1212 18:39:13.218620 2737 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.218891 kubelet[2737]: I1212 18:39:13.218864 2737 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.219012 kubelet[2737]: I1212 18:39:13.218682 2737 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.229329 kubelet[2737]: E1212 18:39:13.229284 2737 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" already exists" pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.230987 kubelet[2737]: E1212 18:39:13.230930 2737 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459-2-2-7-adcd881523\" already exists" pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.277047 kubelet[2737]: I1212 18:39:13.277002 2737 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.286450 kubelet[2737]: I1212 18:39:13.286365 2737 kubelet_node_status.go:124] "Node was previously registered" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.286642 kubelet[2737]: I1212 18:39:13.286484 2737 kubelet_node_status.go:78] "Successfully registered node" node="ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315367 kubelet[2737]: I1212 18:39:13.315334 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c5a27f6c1bcf716779a9e59fcfc8c831-ca-certs\") pod \"kube-apiserver-ci-4459-2-2-7-adcd881523\" (UID: \"c5a27f6c1bcf716779a9e59fcfc8c831\") " pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315367 kubelet[2737]: I1212 18:39:13.315366 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c5a27f6c1bcf716779a9e59fcfc8c831-k8s-certs\") pod \"kube-apiserver-ci-4459-2-2-7-adcd881523\" (UID: \"c5a27f6c1bcf716779a9e59fcfc8c831\") " pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315595 kubelet[2737]: I1212 18:39:13.315382 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c5a27f6c1bcf716779a9e59fcfc8c831-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459-2-2-7-adcd881523\" (UID: \"c5a27f6c1bcf716779a9e59fcfc8c831\") " pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315595 kubelet[2737]: I1212 18:39:13.315401 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-flexvolume-dir\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315595 kubelet[2737]: I1212 18:39:13.315434 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-kubeconfig\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315595 kubelet[2737]: I1212 18:39:13.315450 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-ca-certs\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315595 kubelet[2737]: I1212 18:39:13.315463 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-k8s-certs\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315734 kubelet[2737]: I1212 18:39:13.315476 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/dc0e16563336e6225c3aecdefa7940f4-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459-2-2-7-adcd881523\" (UID: \"dc0e16563336e6225c3aecdefa7940f4\") " pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" Dec 12 18:39:13.315734 kubelet[2737]: I1212 18:39:13.315491 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ecd9d4e6c082fd45c85023240c932785-kubeconfig\") pod \"kube-scheduler-ci-4459-2-2-7-adcd881523\" (UID: \"ecd9d4e6c082fd45c85023240c932785\") " pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:14.086913 kubelet[2737]: I1212 18:39:14.084825 2737 apiserver.go:52] "Watching apiserver" Dec 12 18:39:14.115042 kubelet[2737]: I1212 18:39:14.114311 2737 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Dec 12 18:39:14.155237 kubelet[2737]: I1212 18:39:14.152540 2737 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:14.155519 kubelet[2737]: I1212 18:39:14.155497 2737 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:14.175923 kubelet[2737]: E1212 18:39:14.175884 2737 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459-2-2-7-adcd881523\" already exists" pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" Dec 12 18:39:14.176389 kubelet[2737]: E1212 18:39:14.176250 2737 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459-2-2-7-adcd881523\" already exists" pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" Dec 12 18:39:14.214517 kubelet[2737]: I1212 18:39:14.214446 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4459-2-2-7-adcd881523" podStartSLOduration=3.214428659 podStartE2EDuration="3.214428659s" podCreationTimestamp="2025-12-12 18:39:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:39:14.200480232 +0000 UTC m=+1.179326872" watchObservedRunningTime="2025-12-12 18:39:14.214428659 +0000 UTC m=+1.193275279" Dec 12 18:39:14.228674 kubelet[2737]: I1212 18:39:14.228328 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4459-2-2-7-adcd881523" podStartSLOduration=1.228308549 podStartE2EDuration="1.228308549s" podCreationTimestamp="2025-12-12 18:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:39:14.215743756 +0000 UTC m=+1.194590376" watchObservedRunningTime="2025-12-12 18:39:14.228308549 +0000 UTC m=+1.207155159" Dec 12 18:39:14.247872 kubelet[2737]: I1212 18:39:14.247208 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4459-2-2-7-adcd881523" podStartSLOduration=2.2471653209999998 podStartE2EDuration="2.247165321s" podCreationTimestamp="2025-12-12 18:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:39:14.229085018 +0000 UTC m=+1.207931628" watchObservedRunningTime="2025-12-12 18:39:14.247165321 +0000 UTC m=+1.226011931" Dec 12 18:39:18.268616 kubelet[2737]: I1212 18:39:18.268502 2737 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Dec 12 18:39:18.269670 containerd[1549]: time="2025-12-12T18:39:18.269646115Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 12 18:39:18.270700 kubelet[2737]: I1212 18:39:18.270278 2737 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Dec 12 18:39:19.162758 update_engine[1524]: I20251212 18:39:19.162250 1524 update_attempter.cc:509] Updating boot flags... Dec 12 18:39:19.358696 kubelet[2737]: I1212 18:39:19.358295 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/c538ef43-a25b-4c26-9734-c2dbac4a9762-kube-proxy\") pod \"kube-proxy-2x6dm\" (UID: \"c538ef43-a25b-4c26-9734-c2dbac4a9762\") " pod="kube-system/kube-proxy-2x6dm" Dec 12 18:39:19.358696 kubelet[2737]: I1212 18:39:19.358342 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c538ef43-a25b-4c26-9734-c2dbac4a9762-xtables-lock\") pod \"kube-proxy-2x6dm\" (UID: \"c538ef43-a25b-4c26-9734-c2dbac4a9762\") " pod="kube-system/kube-proxy-2x6dm" Dec 12 18:39:19.358696 kubelet[2737]: I1212 18:39:19.358362 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c538ef43-a25b-4c26-9734-c2dbac4a9762-lib-modules\") pod \"kube-proxy-2x6dm\" (UID: \"c538ef43-a25b-4c26-9734-c2dbac4a9762\") " pod="kube-system/kube-proxy-2x6dm" Dec 12 18:39:19.358696 kubelet[2737]: I1212 18:39:19.358376 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqkdq\" (UniqueName: \"kubernetes.io/projected/c538ef43-a25b-4c26-9734-c2dbac4a9762-kube-api-access-jqkdq\") pod \"kube-proxy-2x6dm\" (UID: \"c538ef43-a25b-4c26-9734-c2dbac4a9762\") " pod="kube-system/kube-proxy-2x6dm" Dec 12 18:39:19.396590 systemd[1]: Created slice kubepods-besteffort-podc538ef43_a25b_4c26_9734_c2dbac4a9762.slice - libcontainer container kubepods-besteffort-podc538ef43_a25b_4c26_9734_c2dbac4a9762.slice. Dec 12 18:39:19.446275 systemd[1]: Created slice kubepods-besteffort-podb1abe97d_df12_46a5_995a_560d250ab0e4.slice - libcontainer container kubepods-besteffort-podb1abe97d_df12_46a5_995a_560d250ab0e4.slice. Dec 12 18:39:19.458586 kubelet[2737]: I1212 18:39:19.458557 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csrfq\" (UniqueName: \"kubernetes.io/projected/b1abe97d-df12-46a5-995a-560d250ab0e4-kube-api-access-csrfq\") pod \"tigera-operator-7dcd859c48-msrgp\" (UID: \"b1abe97d-df12-46a5-995a-560d250ab0e4\") " pod="tigera-operator/tigera-operator-7dcd859c48-msrgp" Dec 12 18:39:19.458666 kubelet[2737]: I1212 18:39:19.458591 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/b1abe97d-df12-46a5-995a-560d250ab0e4-var-lib-calico\") pod \"tigera-operator-7dcd859c48-msrgp\" (UID: \"b1abe97d-df12-46a5-995a-560d250ab0e4\") " pod="tigera-operator/tigera-operator-7dcd859c48-msrgp" Dec 12 18:39:19.509099 kubelet[2737]: E1212 18:39:19.509066 2737 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Dec 12 18:39:19.509099 kubelet[2737]: E1212 18:39:19.509103 2737 projected.go:194] Error preparing data for projected volume kube-api-access-jqkdq for pod kube-system/kube-proxy-2x6dm: configmap "kube-root-ca.crt" not found Dec 12 18:39:19.509248 kubelet[2737]: E1212 18:39:19.509165 2737 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c538ef43-a25b-4c26-9734-c2dbac4a9762-kube-api-access-jqkdq podName:c538ef43-a25b-4c26-9734-c2dbac4a9762 nodeName:}" failed. No retries permitted until 2025-12-12 18:39:20.009141547 +0000 UTC m=+6.987988156 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-jqkdq" (UniqueName: "kubernetes.io/projected/c538ef43-a25b-4c26-9734-c2dbac4a9762-kube-api-access-jqkdq") pod "kube-proxy-2x6dm" (UID: "c538ef43-a25b-4c26-9734-c2dbac4a9762") : configmap "kube-root-ca.crt" not found Dec 12 18:39:19.751685 containerd[1549]: time="2025-12-12T18:39:19.751632706Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-msrgp,Uid:b1abe97d-df12-46a5-995a-560d250ab0e4,Namespace:tigera-operator,Attempt:0,}" Dec 12 18:39:19.771143 containerd[1549]: time="2025-12-12T18:39:19.771017545Z" level=info msg="connecting to shim 7c0c7d6252009e6c6197321aa94386e40d1e0ca3a027b91d8fd4f01b3437e493" address="unix:///run/containerd/s/db50f5a177ed5e66ca7f65cb3785c69326ba1a2243567214f51648e7176a626b" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:19.812374 systemd[1]: Started cri-containerd-7c0c7d6252009e6c6197321aa94386e40d1e0ca3a027b91d8fd4f01b3437e493.scope - libcontainer container 7c0c7d6252009e6c6197321aa94386e40d1e0ca3a027b91d8fd4f01b3437e493. Dec 12 18:39:19.861637 containerd[1549]: time="2025-12-12T18:39:19.861582627Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-msrgp,Uid:b1abe97d-df12-46a5-995a-560d250ab0e4,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"7c0c7d6252009e6c6197321aa94386e40d1e0ca3a027b91d8fd4f01b3437e493\"" Dec 12 18:39:19.863331 containerd[1549]: time="2025-12-12T18:39:19.863240399Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Dec 12 18:39:20.312624 containerd[1549]: time="2025-12-12T18:39:20.312581709Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-2x6dm,Uid:c538ef43-a25b-4c26-9734-c2dbac4a9762,Namespace:kube-system,Attempt:0,}" Dec 12 18:39:20.331207 containerd[1549]: time="2025-12-12T18:39:20.329909541Z" level=info msg="connecting to shim 397705fe2e279d66672862bec1789d12a51d3531a12c48560ef03d6c5501d8ac" address="unix:///run/containerd/s/e168f167119f29e6118ed5f5013678ba0dbbe7044e873aa7b85dc87cfd4ec2d3" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:20.351354 systemd[1]: Started cri-containerd-397705fe2e279d66672862bec1789d12a51d3531a12c48560ef03d6c5501d8ac.scope - libcontainer container 397705fe2e279d66672862bec1789d12a51d3531a12c48560ef03d6c5501d8ac. Dec 12 18:39:20.375845 containerd[1549]: time="2025-12-12T18:39:20.375792088Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-2x6dm,Uid:c538ef43-a25b-4c26-9734-c2dbac4a9762,Namespace:kube-system,Attempt:0,} returns sandbox id \"397705fe2e279d66672862bec1789d12a51d3531a12c48560ef03d6c5501d8ac\"" Dec 12 18:39:20.380862 containerd[1549]: time="2025-12-12T18:39:20.380812886Z" level=info msg="CreateContainer within sandbox \"397705fe2e279d66672862bec1789d12a51d3531a12c48560ef03d6c5501d8ac\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 12 18:39:20.390849 containerd[1549]: time="2025-12-12T18:39:20.390804652Z" level=info msg="Container 3a036c7666d7f80ca36f395de9363e3b1a2176f6609922ed9b5cd74bd8b7d001: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:20.399525 containerd[1549]: time="2025-12-12T18:39:20.399391546Z" level=info msg="CreateContainer within sandbox \"397705fe2e279d66672862bec1789d12a51d3531a12c48560ef03d6c5501d8ac\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"3a036c7666d7f80ca36f395de9363e3b1a2176f6609922ed9b5cd74bd8b7d001\"" Dec 12 18:39:20.402172 containerd[1549]: time="2025-12-12T18:39:20.402151449Z" level=info msg="StartContainer for \"3a036c7666d7f80ca36f395de9363e3b1a2176f6609922ed9b5cd74bd8b7d001\"" Dec 12 18:39:20.405395 containerd[1549]: time="2025-12-12T18:39:20.404483948Z" level=info msg="connecting to shim 3a036c7666d7f80ca36f395de9363e3b1a2176f6609922ed9b5cd74bd8b7d001" address="unix:///run/containerd/s/e168f167119f29e6118ed5f5013678ba0dbbe7044e873aa7b85dc87cfd4ec2d3" protocol=ttrpc version=3 Dec 12 18:39:20.422340 systemd[1]: Started cri-containerd-3a036c7666d7f80ca36f395de9363e3b1a2176f6609922ed9b5cd74bd8b7d001.scope - libcontainer container 3a036c7666d7f80ca36f395de9363e3b1a2176f6609922ed9b5cd74bd8b7d001. Dec 12 18:39:20.477838 containerd[1549]: time="2025-12-12T18:39:20.477797450Z" level=info msg="StartContainer for \"3a036c7666d7f80ca36f395de9363e3b1a2176f6609922ed9b5cd74bd8b7d001\" returns successfully" Dec 12 18:39:21.185144 kubelet[2737]: I1212 18:39:21.185044 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-2x6dm" podStartSLOduration=2.185020691 podStartE2EDuration="2.185020691s" podCreationTimestamp="2025-12-12 18:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:39:21.184749005 +0000 UTC m=+8.163595646" watchObservedRunningTime="2025-12-12 18:39:21.185020691 +0000 UTC m=+8.163867331" Dec 12 18:39:21.875350 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3151377420.mount: Deactivated successfully. Dec 12 18:39:22.469876 containerd[1549]: time="2025-12-12T18:39:22.469799408Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:22.470928 containerd[1549]: time="2025-12-12T18:39:22.470831380Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=25061691" Dec 12 18:39:22.471832 containerd[1549]: time="2025-12-12T18:39:22.471797368Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:22.474224 containerd[1549]: time="2025-12-12T18:39:22.473914951Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:22.474793 containerd[1549]: time="2025-12-12T18:39:22.474569880Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 2.611270782s" Dec 12 18:39:22.474793 containerd[1549]: time="2025-12-12T18:39:22.474603183Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Dec 12 18:39:22.478918 containerd[1549]: time="2025-12-12T18:39:22.478863635Z" level=info msg="CreateContainer within sandbox \"7c0c7d6252009e6c6197321aa94386e40d1e0ca3a027b91d8fd4f01b3437e493\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Dec 12 18:39:22.484618 containerd[1549]: time="2025-12-12T18:39:22.484510108Z" level=info msg="Container f7acc4d438d27ce55ea55979c6074c3918cbd4ba68b8d45033764e022dfd4e0f: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:22.494374 containerd[1549]: time="2025-12-12T18:39:22.494320996Z" level=info msg="CreateContainer within sandbox \"7c0c7d6252009e6c6197321aa94386e40d1e0ca3a027b91d8fd4f01b3437e493\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"f7acc4d438d27ce55ea55979c6074c3918cbd4ba68b8d45033764e022dfd4e0f\"" Dec 12 18:39:22.495759 containerd[1549]: time="2025-12-12T18:39:22.495133679Z" level=info msg="StartContainer for \"f7acc4d438d27ce55ea55979c6074c3918cbd4ba68b8d45033764e022dfd4e0f\"" Dec 12 18:39:22.496907 containerd[1549]: time="2025-12-12T18:39:22.496764746Z" level=info msg="connecting to shim f7acc4d438d27ce55ea55979c6074c3918cbd4ba68b8d45033764e022dfd4e0f" address="unix:///run/containerd/s/db50f5a177ed5e66ca7f65cb3785c69326ba1a2243567214f51648e7176a626b" protocol=ttrpc version=3 Dec 12 18:39:22.522330 systemd[1]: Started cri-containerd-f7acc4d438d27ce55ea55979c6074c3918cbd4ba68b8d45033764e022dfd4e0f.scope - libcontainer container f7acc4d438d27ce55ea55979c6074c3918cbd4ba68b8d45033764e022dfd4e0f. Dec 12 18:39:22.552038 containerd[1549]: time="2025-12-12T18:39:22.551999619Z" level=info msg="StartContainer for \"f7acc4d438d27ce55ea55979c6074c3918cbd4ba68b8d45033764e022dfd4e0f\" returns successfully" Dec 12 18:39:23.196500 kubelet[2737]: I1212 18:39:23.196429 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-msrgp" podStartSLOduration=1.5836828569999999 podStartE2EDuration="4.196409639s" podCreationTimestamp="2025-12-12 18:39:19 +0000 UTC" firstStartedPulling="2025-12-12 18:39:19.86286425 +0000 UTC m=+6.841710870" lastFinishedPulling="2025-12-12 18:39:22.475591042 +0000 UTC m=+9.454437652" observedRunningTime="2025-12-12 18:39:23.195978415 +0000 UTC m=+10.174825136" watchObservedRunningTime="2025-12-12 18:39:23.196409639 +0000 UTC m=+10.175256289" Dec 12 18:39:28.603302 sudo[1806]: pam_unix(sudo:session): session closed for user root Dec 12 18:39:28.763002 sshd[1805]: Connection closed by 147.75.109.163 port 58180 Dec 12 18:39:28.763412 sshd-session[1802]: pam_unix(sshd:session): session closed for user core Dec 12 18:39:28.766431 systemd[1]: sshd@6-65.21.248.196:22-147.75.109.163:58180.service: Deactivated successfully. Dec 12 18:39:28.769002 systemd[1]: session-7.scope: Deactivated successfully. Dec 12 18:39:28.770354 systemd[1]: session-7.scope: Consumed 4.869s CPU time, 159.8M memory peak. Dec 12 18:39:28.774899 systemd-logind[1520]: Session 7 logged out. Waiting for processes to exit. Dec 12 18:39:28.776583 systemd-logind[1520]: Removed session 7. Dec 12 18:39:32.848252 systemd[1]: Created slice kubepods-besteffort-pod58e4905b_48ee_4231_81a4_86b910ae3ae6.slice - libcontainer container kubepods-besteffort-pod58e4905b_48ee_4231_81a4_86b910ae3ae6.slice. Dec 12 18:39:32.848623 kubelet[2737]: I1212 18:39:32.848603 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/58e4905b-48ee-4231-81a4-86b910ae3ae6-typha-certs\") pod \"calico-typha-54fbdf5767-6h8rf\" (UID: \"58e4905b-48ee-4231-81a4-86b910ae3ae6\") " pod="calico-system/calico-typha-54fbdf5767-6h8rf" Dec 12 18:39:32.849082 kubelet[2737]: I1212 18:39:32.848628 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58e4905b-48ee-4231-81a4-86b910ae3ae6-tigera-ca-bundle\") pod \"calico-typha-54fbdf5767-6h8rf\" (UID: \"58e4905b-48ee-4231-81a4-86b910ae3ae6\") " pod="calico-system/calico-typha-54fbdf5767-6h8rf" Dec 12 18:39:32.849082 kubelet[2737]: I1212 18:39:32.848643 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcpgz\" (UniqueName: \"kubernetes.io/projected/58e4905b-48ee-4231-81a4-86b910ae3ae6-kube-api-access-kcpgz\") pod \"calico-typha-54fbdf5767-6h8rf\" (UID: \"58e4905b-48ee-4231-81a4-86b910ae3ae6\") " pod="calico-system/calico-typha-54fbdf5767-6h8rf" Dec 12 18:39:33.035869 systemd[1]: Created slice kubepods-besteffort-pod2f174a49_0a20_48cd_8464_f19ee56ddcdf.slice - libcontainer container kubepods-besteffort-pod2f174a49_0a20_48cd_8464_f19ee56ddcdf.slice. Dec 12 18:39:33.050827 kubelet[2737]: I1212 18:39:33.050781 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-cni-log-dir\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.050947 kubelet[2737]: I1212 18:39:33.050872 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-xtables-lock\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.050947 kubelet[2737]: I1212 18:39:33.050893 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-var-lib-calico\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.050947 kubelet[2737]: I1212 18:39:33.050904 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-cni-net-dir\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.051004 kubelet[2737]: I1212 18:39:33.050947 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-lib-modules\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.051004 kubelet[2737]: I1212 18:39:33.050962 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-policysync\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.051392 kubelet[2737]: I1212 18:39:33.051036 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-cni-bin-dir\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.051392 kubelet[2737]: I1212 18:39:33.051107 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-flexvol-driver-host\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.051392 kubelet[2737]: I1212 18:39:33.051126 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f174a49-0a20-48cd-8464-f19ee56ddcdf-tigera-ca-bundle\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.051392 kubelet[2737]: I1212 18:39:33.051137 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/2f174a49-0a20-48cd-8464-f19ee56ddcdf-var-run-calico\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.051392 kubelet[2737]: I1212 18:39:33.051176 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/2f174a49-0a20-48cd-8464-f19ee56ddcdf-node-certs\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.051538 kubelet[2737]: I1212 18:39:33.051216 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcsg9\" (UniqueName: \"kubernetes.io/projected/2f174a49-0a20-48cd-8464-f19ee56ddcdf-kube-api-access-gcsg9\") pod \"calico-node-ht5sj\" (UID: \"2f174a49-0a20-48cd-8464-f19ee56ddcdf\") " pod="calico-system/calico-node-ht5sj" Dec 12 18:39:33.154006 containerd[1549]: time="2025-12-12T18:39:33.153441153Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-54fbdf5767-6h8rf,Uid:58e4905b-48ee-4231-81a4-86b910ae3ae6,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:33.160715 kubelet[2737]: E1212 18:39:33.160292 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.160715 kubelet[2737]: W1212 18:39:33.160449 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.160715 kubelet[2737]: E1212 18:39:33.160482 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.164255 kubelet[2737]: E1212 18:39:33.161932 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.164255 kubelet[2737]: W1212 18:39:33.161958 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.164255 kubelet[2737]: E1212 18:39:33.162225 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.164255 kubelet[2737]: E1212 18:39:33.163297 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.164255 kubelet[2737]: W1212 18:39:33.163570 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.164255 kubelet[2737]: E1212 18:39:33.163599 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.166108 kubelet[2737]: E1212 18:39:33.166070 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.166108 kubelet[2737]: W1212 18:39:33.166107 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.166284 kubelet[2737]: E1212 18:39:33.166127 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.171541 kubelet[2737]: E1212 18:39:33.166494 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.171541 kubelet[2737]: W1212 18:39:33.166514 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.171541 kubelet[2737]: E1212 18:39:33.166532 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.171541 kubelet[2737]: E1212 18:39:33.168533 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.171541 kubelet[2737]: W1212 18:39:33.168549 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.171541 kubelet[2737]: E1212 18:39:33.168566 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.171541 kubelet[2737]: E1212 18:39:33.168768 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.171541 kubelet[2737]: W1212 18:39:33.168781 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.171541 kubelet[2737]: E1212 18:39:33.168797 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.171541 kubelet[2737]: E1212 18:39:33.170158 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.172052 kubelet[2737]: W1212 18:39:33.170174 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.172772 kubelet[2737]: E1212 18:39:33.172275 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.172912 kubelet[2737]: E1212 18:39:33.172783 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.172912 kubelet[2737]: W1212 18:39:33.172808 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.172912 kubelet[2737]: E1212 18:39:33.172848 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.175153 kubelet[2737]: E1212 18:39:33.175031 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.175153 kubelet[2737]: W1212 18:39:33.175056 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.175153 kubelet[2737]: E1212 18:39:33.175073 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.192184 kubelet[2737]: E1212 18:39:33.188007 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.192184 kubelet[2737]: W1212 18:39:33.188032 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.192184 kubelet[2737]: E1212 18:39:33.188055 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.271548 containerd[1549]: time="2025-12-12T18:39:33.271482377Z" level=info msg="connecting to shim f8dad181068e7329f0c8d948ea3a10cb2a6edeb6a8fa6bb18bc7816ae91f75e8" address="unix:///run/containerd/s/58f24266e6e3bb1b632b18b7411aadc2d202de4b637c93b7f7cd30e064ca0ee4" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:33.301561 kubelet[2737]: E1212 18:39:33.301519 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:39:33.311608 systemd[1]: Started cri-containerd-f8dad181068e7329f0c8d948ea3a10cb2a6edeb6a8fa6bb18bc7816ae91f75e8.scope - libcontainer container f8dad181068e7329f0c8d948ea3a10cb2a6edeb6a8fa6bb18bc7816ae91f75e8. Dec 12 18:39:33.342263 containerd[1549]: time="2025-12-12T18:39:33.342217986Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ht5sj,Uid:2f174a49-0a20-48cd-8464-f19ee56ddcdf,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:33.350514 kubelet[2737]: E1212 18:39:33.350486 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.350514 kubelet[2737]: W1212 18:39:33.350509 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.350646 kubelet[2737]: E1212 18:39:33.350529 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.351208 kubelet[2737]: E1212 18:39:33.351180 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.351325 kubelet[2737]: W1212 18:39:33.351311 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.351409 kubelet[2737]: E1212 18:39:33.351400 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.351686 kubelet[2737]: E1212 18:39:33.351637 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.351686 kubelet[2737]: W1212 18:39:33.351649 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.351686 kubelet[2737]: E1212 18:39:33.351662 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.352794 kubelet[2737]: E1212 18:39:33.352742 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.352794 kubelet[2737]: W1212 18:39:33.352751 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.352794 kubelet[2737]: E1212 18:39:33.352761 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.355445 kubelet[2737]: E1212 18:39:33.355432 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.355445 kubelet[2737]: W1212 18:39:33.355443 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.355513 kubelet[2737]: E1212 18:39:33.355454 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.355924 kubelet[2737]: E1212 18:39:33.355909 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.355924 kubelet[2737]: W1212 18:39:33.355923 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.356024 kubelet[2737]: E1212 18:39:33.355932 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.356315 kubelet[2737]: E1212 18:39:33.356267 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.356315 kubelet[2737]: W1212 18:39:33.356281 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.356315 kubelet[2737]: E1212 18:39:33.356289 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.357271 kubelet[2737]: E1212 18:39:33.357255 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.357271 kubelet[2737]: W1212 18:39:33.357269 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.357271 kubelet[2737]: E1212 18:39:33.357278 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.357482 kubelet[2737]: E1212 18:39:33.357415 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.357482 kubelet[2737]: W1212 18:39:33.357421 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.357482 kubelet[2737]: E1212 18:39:33.357428 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.357655 kubelet[2737]: E1212 18:39:33.357634 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.357655 kubelet[2737]: W1212 18:39:33.357646 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.357655 kubelet[2737]: E1212 18:39:33.357653 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.357919 kubelet[2737]: E1212 18:39:33.357845 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.357919 kubelet[2737]: W1212 18:39:33.357855 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.357919 kubelet[2737]: E1212 18:39:33.357864 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.358274 kubelet[2737]: E1212 18:39:33.358202 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.358274 kubelet[2737]: W1212 18:39:33.358212 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.358274 kubelet[2737]: E1212 18:39:33.358220 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.358685 kubelet[2737]: E1212 18:39:33.358438 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.358685 kubelet[2737]: W1212 18:39:33.358446 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.358685 kubelet[2737]: E1212 18:39:33.358453 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.359439 kubelet[2737]: E1212 18:39:33.359222 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.359439 kubelet[2737]: W1212 18:39:33.359232 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.359439 kubelet[2737]: E1212 18:39:33.359240 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.359439 kubelet[2737]: E1212 18:39:33.359344 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.359439 kubelet[2737]: W1212 18:39:33.359355 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.359439 kubelet[2737]: E1212 18:39:33.359367 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.360325 kubelet[2737]: E1212 18:39:33.360241 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.360325 kubelet[2737]: W1212 18:39:33.360250 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.360325 kubelet[2737]: E1212 18:39:33.360258 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.360618 kubelet[2737]: E1212 18:39:33.360504 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.360618 kubelet[2737]: W1212 18:39:33.360517 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.360618 kubelet[2737]: E1212 18:39:33.360531 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.360880 kubelet[2737]: E1212 18:39:33.360800 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.360880 kubelet[2737]: W1212 18:39:33.360824 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.360880 kubelet[2737]: E1212 18:39:33.360832 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.361016 kubelet[2737]: E1212 18:39:33.361007 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.361140 kubelet[2737]: W1212 18:39:33.361060 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.361140 kubelet[2737]: E1212 18:39:33.361070 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.361511 kubelet[2737]: E1212 18:39:33.361501 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.361642 kubelet[2737]: W1212 18:39:33.361567 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.361860 kubelet[2737]: E1212 18:39:33.361699 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.362465 kubelet[2737]: E1212 18:39:33.362455 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.362737 kubelet[2737]: W1212 18:39:33.362721 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.362801 kubelet[2737]: E1212 18:39:33.362793 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.362888 kubelet[2737]: I1212 18:39:33.362877 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cqz8\" (UniqueName: \"kubernetes.io/projected/583f364c-28f5-44f4-853f-72900703e8a2-kube-api-access-9cqz8\") pod \"csi-node-driver-2msts\" (UID: \"583f364c-28f5-44f4-853f-72900703e8a2\") " pod="calico-system/csi-node-driver-2msts" Dec 12 18:39:33.363111 kubelet[2737]: E1212 18:39:33.363031 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.363111 kubelet[2737]: W1212 18:39:33.363040 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.363111 kubelet[2737]: E1212 18:39:33.363048 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.363111 kubelet[2737]: I1212 18:39:33.363079 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/583f364c-28f5-44f4-853f-72900703e8a2-registration-dir\") pod \"csi-node-driver-2msts\" (UID: \"583f364c-28f5-44f4-853f-72900703e8a2\") " pod="calico-system/csi-node-driver-2msts" Dec 12 18:39:33.364293 kubelet[2737]: E1212 18:39:33.364248 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.364293 kubelet[2737]: W1212 18:39:33.364259 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.364293 kubelet[2737]: E1212 18:39:33.364270 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.364497 kubelet[2737]: E1212 18:39:33.364454 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.364497 kubelet[2737]: W1212 18:39:33.364464 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.364497 kubelet[2737]: E1212 18:39:33.364472 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.364726 kubelet[2737]: E1212 18:39:33.364715 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.364829 kubelet[2737]: W1212 18:39:33.364782 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.364829 kubelet[2737]: E1212 18:39:33.364803 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.364994 kubelet[2737]: I1212 18:39:33.364969 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/583f364c-28f5-44f4-853f-72900703e8a2-kubelet-dir\") pod \"csi-node-driver-2msts\" (UID: \"583f364c-28f5-44f4-853f-72900703e8a2\") " pod="calico-system/csi-node-driver-2msts" Dec 12 18:39:33.365155 kubelet[2737]: E1212 18:39:33.365123 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.365155 kubelet[2737]: W1212 18:39:33.365137 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.365155 kubelet[2737]: E1212 18:39:33.365145 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.365406 kubelet[2737]: E1212 18:39:33.365383 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.365406 kubelet[2737]: W1212 18:39:33.365391 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.365406 kubelet[2737]: E1212 18:39:33.365398 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.365634 kubelet[2737]: E1212 18:39:33.365610 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.365634 kubelet[2737]: W1212 18:39:33.365618 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.365634 kubelet[2737]: E1212 18:39:33.365626 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.365777 kubelet[2737]: I1212 18:39:33.365734 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/583f364c-28f5-44f4-853f-72900703e8a2-socket-dir\") pod \"csi-node-driver-2msts\" (UID: \"583f364c-28f5-44f4-853f-72900703e8a2\") " pod="calico-system/csi-node-driver-2msts" Dec 12 18:39:33.365980 kubelet[2737]: E1212 18:39:33.365946 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.365980 kubelet[2737]: W1212 18:39:33.365963 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.365980 kubelet[2737]: E1212 18:39:33.365971 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.366127 kubelet[2737]: I1212 18:39:33.366077 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/583f364c-28f5-44f4-853f-72900703e8a2-varrun\") pod \"csi-node-driver-2msts\" (UID: \"583f364c-28f5-44f4-853f-72900703e8a2\") " pod="calico-system/csi-node-driver-2msts" Dec 12 18:39:33.366333 kubelet[2737]: E1212 18:39:33.366308 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.366333 kubelet[2737]: W1212 18:39:33.366317 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.366333 kubelet[2737]: E1212 18:39:33.366324 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.366581 kubelet[2737]: E1212 18:39:33.366557 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.366581 kubelet[2737]: W1212 18:39:33.366566 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.366581 kubelet[2737]: E1212 18:39:33.366573 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.366885 kubelet[2737]: E1212 18:39:33.366855 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.366885 kubelet[2737]: W1212 18:39:33.366864 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.366885 kubelet[2737]: E1212 18:39:33.366871 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.367110 kubelet[2737]: E1212 18:39:33.367085 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.367110 kubelet[2737]: W1212 18:39:33.367094 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.367110 kubelet[2737]: E1212 18:39:33.367101 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.367334 kubelet[2737]: E1212 18:39:33.367309 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.367334 kubelet[2737]: W1212 18:39:33.367317 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.367334 kubelet[2737]: E1212 18:39:33.367323 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.367557 kubelet[2737]: E1212 18:39:33.367530 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.367557 kubelet[2737]: W1212 18:39:33.367538 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.367557 kubelet[2737]: E1212 18:39:33.367544 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.369890 containerd[1549]: time="2025-12-12T18:39:33.369831956Z" level=info msg="connecting to shim 508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277" address="unix:///run/containerd/s/343d31595ab3d1ab90348b49fd8f71e383b8f1e0e2ab4a166558feae6c86541d" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:33.386539 containerd[1549]: time="2025-12-12T18:39:33.386487599Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-54fbdf5767-6h8rf,Uid:58e4905b-48ee-4231-81a4-86b910ae3ae6,Namespace:calico-system,Attempt:0,} returns sandbox id \"f8dad181068e7329f0c8d948ea3a10cb2a6edeb6a8fa6bb18bc7816ae91f75e8\"" Dec 12 18:39:33.394545 containerd[1549]: time="2025-12-12T18:39:33.394329714Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Dec 12 18:39:33.435327 systemd[1]: Started cri-containerd-508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277.scope - libcontainer container 508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277. Dec 12 18:39:33.467064 kubelet[2737]: E1212 18:39:33.467031 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.467064 kubelet[2737]: W1212 18:39:33.467056 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.467181 kubelet[2737]: E1212 18:39:33.467077 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.467507 containerd[1549]: time="2025-12-12T18:39:33.467344660Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ht5sj,Uid:2f174a49-0a20-48cd-8464-f19ee56ddcdf,Namespace:calico-system,Attempt:0,} returns sandbox id \"508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277\"" Dec 12 18:39:33.467614 kubelet[2737]: E1212 18:39:33.467593 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.467746 kubelet[2737]: W1212 18:39:33.467725 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.467778 kubelet[2737]: E1212 18:39:33.467747 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.468112 kubelet[2737]: E1212 18:39:33.468094 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.468112 kubelet[2737]: W1212 18:39:33.468108 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.468176 kubelet[2737]: E1212 18:39:33.468120 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.468767 kubelet[2737]: E1212 18:39:33.468735 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.468767 kubelet[2737]: W1212 18:39:33.468763 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.468838 kubelet[2737]: E1212 18:39:33.468775 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.469276 kubelet[2737]: E1212 18:39:33.469256 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.469276 kubelet[2737]: W1212 18:39:33.469271 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.469347 kubelet[2737]: E1212 18:39:33.469285 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.469999 kubelet[2737]: E1212 18:39:33.469980 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.469999 kubelet[2737]: W1212 18:39:33.469994 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.469999 kubelet[2737]: E1212 18:39:33.470005 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.470405 kubelet[2737]: E1212 18:39:33.470390 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.470405 kubelet[2737]: W1212 18:39:33.470403 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.470461 kubelet[2737]: E1212 18:39:33.470414 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.470874 kubelet[2737]: E1212 18:39:33.470839 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.470874 kubelet[2737]: W1212 18:39:33.470852 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.470874 kubelet[2737]: E1212 18:39:33.470860 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.471400 kubelet[2737]: E1212 18:39:33.471378 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.471400 kubelet[2737]: W1212 18:39:33.471397 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.471482 kubelet[2737]: E1212 18:39:33.471413 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.472295 kubelet[2737]: E1212 18:39:33.472278 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.472295 kubelet[2737]: W1212 18:39:33.472290 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.472407 kubelet[2737]: E1212 18:39:33.472300 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.472432 kubelet[2737]: E1212 18:39:33.472413 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.472432 kubelet[2737]: W1212 18:39:33.472419 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.472467 kubelet[2737]: E1212 18:39:33.472444 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.472577 kubelet[2737]: E1212 18:39:33.472569 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.472577 kubelet[2737]: W1212 18:39:33.472577 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.472632 kubelet[2737]: E1212 18:39:33.472584 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.472986 kubelet[2737]: E1212 18:39:33.472955 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.472986 kubelet[2737]: W1212 18:39:33.472972 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.472986 kubelet[2737]: E1212 18:39:33.472984 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.473281 kubelet[2737]: E1212 18:39:33.473260 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.473281 kubelet[2737]: W1212 18:39:33.473279 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.473281 kubelet[2737]: E1212 18:39:33.473287 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.473424 kubelet[2737]: E1212 18:39:33.473403 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.473424 kubelet[2737]: W1212 18:39:33.473416 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.473424 kubelet[2737]: E1212 18:39:33.473423 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.473580 kubelet[2737]: E1212 18:39:33.473563 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.473580 kubelet[2737]: W1212 18:39:33.473573 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.473580 kubelet[2737]: E1212 18:39:33.473580 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.473719 kubelet[2737]: E1212 18:39:33.473700 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.473719 kubelet[2737]: W1212 18:39:33.473711 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.473719 kubelet[2737]: E1212 18:39:33.473718 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.473981 kubelet[2737]: E1212 18:39:33.473949 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.473981 kubelet[2737]: W1212 18:39:33.473966 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.473981 kubelet[2737]: E1212 18:39:33.473975 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.474245 kubelet[2737]: E1212 18:39:33.474181 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.474245 kubelet[2737]: W1212 18:39:33.474235 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.474313 kubelet[2737]: E1212 18:39:33.474246 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.474521 kubelet[2737]: E1212 18:39:33.474490 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.474521 kubelet[2737]: W1212 18:39:33.474505 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.474521 kubelet[2737]: E1212 18:39:33.474513 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.474708 kubelet[2737]: E1212 18:39:33.474690 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.474708 kubelet[2737]: W1212 18:39:33.474702 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.474708 kubelet[2737]: E1212 18:39:33.474709 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.474899 kubelet[2737]: E1212 18:39:33.474887 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.474899 kubelet[2737]: W1212 18:39:33.474896 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.474984 kubelet[2737]: E1212 18:39:33.474904 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.475094 kubelet[2737]: E1212 18:39:33.475071 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.475094 kubelet[2737]: W1212 18:39:33.475087 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.475094 kubelet[2737]: E1212 18:39:33.475094 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.475270 kubelet[2737]: E1212 18:39:33.475250 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.475270 kubelet[2737]: W1212 18:39:33.475261 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.475270 kubelet[2737]: E1212 18:39:33.475267 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.475414 kubelet[2737]: E1212 18:39:33.475393 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.475414 kubelet[2737]: W1212 18:39:33.475410 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.475466 kubelet[2737]: E1212 18:39:33.475418 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:33.484310 kubelet[2737]: E1212 18:39:33.484286 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:33.484310 kubelet[2737]: W1212 18:39:33.484302 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:33.484310 kubelet[2737]: E1212 18:39:33.484313 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:35.118658 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2666786851.mount: Deactivated successfully. Dec 12 18:39:35.121637 kubelet[2737]: E1212 18:39:35.121421 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:39:35.613071 containerd[1549]: time="2025-12-12T18:39:35.612991220Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:35.614020 containerd[1549]: time="2025-12-12T18:39:35.613871435Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=35234628" Dec 12 18:39:35.614739 containerd[1549]: time="2025-12-12T18:39:35.614713930Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:35.616360 containerd[1549]: time="2025-12-12T18:39:35.616320794Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:35.616827 containerd[1549]: time="2025-12-12T18:39:35.616796333Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 2.222436383s" Dec 12 18:39:35.616896 containerd[1549]: time="2025-12-12T18:39:35.616884196Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Dec 12 18:39:35.617966 containerd[1549]: time="2025-12-12T18:39:35.617944529Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Dec 12 18:39:35.627450 containerd[1549]: time="2025-12-12T18:39:35.627026297Z" level=info msg="CreateContainer within sandbox \"f8dad181068e7329f0c8d948ea3a10cb2a6edeb6a8fa6bb18bc7816ae91f75e8\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Dec 12 18:39:35.636845 containerd[1549]: time="2025-12-12T18:39:35.636816519Z" level=info msg="Container 25f2f30ea3d34295cb651bea4033e19fe477dfa5cc5fae8796df191b9ce1e785: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:35.651282 containerd[1549]: time="2025-12-12T18:39:35.651242237Z" level=info msg="CreateContainer within sandbox \"f8dad181068e7329f0c8d948ea3a10cb2a6edeb6a8fa6bb18bc7816ae91f75e8\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"25f2f30ea3d34295cb651bea4033e19fe477dfa5cc5fae8796df191b9ce1e785\"" Dec 12 18:39:35.654925 containerd[1549]: time="2025-12-12T18:39:35.654879246Z" level=info msg="StartContainer for \"25f2f30ea3d34295cb651bea4033e19fe477dfa5cc5fae8796df191b9ce1e785\"" Dec 12 18:39:35.656111 containerd[1549]: time="2025-12-12T18:39:35.656076013Z" level=info msg="connecting to shim 25f2f30ea3d34295cb651bea4033e19fe477dfa5cc5fae8796df191b9ce1e785" address="unix:///run/containerd/s/58f24266e6e3bb1b632b18b7411aadc2d202de4b637c93b7f7cd30e064ca0ee4" protocol=ttrpc version=3 Dec 12 18:39:35.671474 systemd[1]: Started cri-containerd-25f2f30ea3d34295cb651bea4033e19fe477dfa5cc5fae8796df191b9ce1e785.scope - libcontainer container 25f2f30ea3d34295cb651bea4033e19fe477dfa5cc5fae8796df191b9ce1e785. Dec 12 18:39:35.719480 containerd[1549]: time="2025-12-12T18:39:35.719334058Z" level=info msg="StartContainer for \"25f2f30ea3d34295cb651bea4033e19fe477dfa5cc5fae8796df191b9ce1e785\" returns successfully" Dec 12 18:39:36.261020 kubelet[2737]: I1212 18:39:36.258123 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-54fbdf5767-6h8rf" podStartSLOduration=2.02894178 podStartE2EDuration="4.258107268s" podCreationTimestamp="2025-12-12 18:39:32 +0000 UTC" firstStartedPulling="2025-12-12 18:39:33.388485823 +0000 UTC m=+20.367332433" lastFinishedPulling="2025-12-12 18:39:35.617651311 +0000 UTC m=+22.596497921" observedRunningTime="2025-12-12 18:39:36.257631539 +0000 UTC m=+23.236478179" watchObservedRunningTime="2025-12-12 18:39:36.258107268 +0000 UTC m=+23.236953888" Dec 12 18:39:36.280177 kubelet[2737]: E1212 18:39:36.280132 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.280177 kubelet[2737]: W1212 18:39:36.280155 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.282523 kubelet[2737]: E1212 18:39:36.282463 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.282891 kubelet[2737]: E1212 18:39:36.282861 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.282891 kubelet[2737]: W1212 18:39:36.282880 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.283113 kubelet[2737]: E1212 18:39:36.282899 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.283113 kubelet[2737]: E1212 18:39:36.283077 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.283113 kubelet[2737]: W1212 18:39:36.283085 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.283113 kubelet[2737]: E1212 18:39:36.283104 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.284020 kubelet[2737]: E1212 18:39:36.283974 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.284020 kubelet[2737]: W1212 18:39:36.283991 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.284020 kubelet[2737]: E1212 18:39:36.284001 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.284480 kubelet[2737]: E1212 18:39:36.284465 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.284480 kubelet[2737]: W1212 18:39:36.284478 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.284662 kubelet[2737]: E1212 18:39:36.284490 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.284662 kubelet[2737]: E1212 18:39:36.284632 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.284662 kubelet[2737]: W1212 18:39:36.284640 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.284662 kubelet[2737]: E1212 18:39:36.284647 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.285042 kubelet[2737]: E1212 18:39:36.285024 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.285042 kubelet[2737]: W1212 18:39:36.285036 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.285042 kubelet[2737]: E1212 18:39:36.285044 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.285365 kubelet[2737]: E1212 18:39:36.285199 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.285365 kubelet[2737]: W1212 18:39:36.285206 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.285365 kubelet[2737]: E1212 18:39:36.285218 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.285479 kubelet[2737]: E1212 18:39:36.285384 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.285479 kubelet[2737]: W1212 18:39:36.285391 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.285479 kubelet[2737]: E1212 18:39:36.285398 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.286356 kubelet[2737]: E1212 18:39:36.286331 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.286356 kubelet[2737]: W1212 18:39:36.286345 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.286356 kubelet[2737]: E1212 18:39:36.286353 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.286989 kubelet[2737]: E1212 18:39:36.286493 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.286989 kubelet[2737]: W1212 18:39:36.286500 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.286989 kubelet[2737]: E1212 18:39:36.286508 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.286989 kubelet[2737]: E1212 18:39:36.286635 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.286989 kubelet[2737]: W1212 18:39:36.286642 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.286989 kubelet[2737]: E1212 18:39:36.286649 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.287301 kubelet[2737]: E1212 18:39:36.287272 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.287301 kubelet[2737]: W1212 18:39:36.287285 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.287301 kubelet[2737]: E1212 18:39:36.287297 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.287881 kubelet[2737]: E1212 18:39:36.287485 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.287881 kubelet[2737]: W1212 18:39:36.287495 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.287881 kubelet[2737]: E1212 18:39:36.287506 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.288325 kubelet[2737]: E1212 18:39:36.288242 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.288325 kubelet[2737]: W1212 18:39:36.288255 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.288325 kubelet[2737]: E1212 18:39:36.288266 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.296054 kubelet[2737]: E1212 18:39:36.296027 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.296054 kubelet[2737]: W1212 18:39:36.296046 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.296054 kubelet[2737]: E1212 18:39:36.296061 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.296455 kubelet[2737]: E1212 18:39:36.296213 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.296455 kubelet[2737]: W1212 18:39:36.296222 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.296455 kubelet[2737]: E1212 18:39:36.296230 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.296836 kubelet[2737]: E1212 18:39:36.296737 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.296836 kubelet[2737]: W1212 18:39:36.296752 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.296836 kubelet[2737]: E1212 18:39:36.296764 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.297264 kubelet[2737]: E1212 18:39:36.297173 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.297264 kubelet[2737]: W1212 18:39:36.297210 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.297264 kubelet[2737]: E1212 18:39:36.297221 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.297879 kubelet[2737]: E1212 18:39:36.297707 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.297879 kubelet[2737]: W1212 18:39:36.297719 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.297879 kubelet[2737]: E1212 18:39:36.297729 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.305302 kubelet[2737]: E1212 18:39:36.305286 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.305461 kubelet[2737]: W1212 18:39:36.305386 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.305461 kubelet[2737]: E1212 18:39:36.305404 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.305852 kubelet[2737]: E1212 18:39:36.305730 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.305852 kubelet[2737]: W1212 18:39:36.305741 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.305852 kubelet[2737]: E1212 18:39:36.305753 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.306003 kubelet[2737]: E1212 18:39:36.305992 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.306074 kubelet[2737]: W1212 18:39:36.306064 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.306212 kubelet[2737]: E1212 18:39:36.306145 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.306435 kubelet[2737]: E1212 18:39:36.306424 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.306596 kubelet[2737]: W1212 18:39:36.306511 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.306596 kubelet[2737]: E1212 18:39:36.306525 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.306874 kubelet[2737]: E1212 18:39:36.306863 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.306959 kubelet[2737]: W1212 18:39:36.306948 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.307462 kubelet[2737]: E1212 18:39:36.307040 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.307462 kubelet[2737]: E1212 18:39:36.307247 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.307462 kubelet[2737]: W1212 18:39:36.307255 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.307462 kubelet[2737]: E1212 18:39:36.307265 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.307703 kubelet[2737]: E1212 18:39:36.307659 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.307703 kubelet[2737]: W1212 18:39:36.307671 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.307882 kubelet[2737]: E1212 18:39:36.307788 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.308095 kubelet[2737]: E1212 18:39:36.308075 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.308209 kubelet[2737]: W1212 18:39:36.308161 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.308209 kubelet[2737]: E1212 18:39:36.308174 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.308639 kubelet[2737]: E1212 18:39:36.308601 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.308639 kubelet[2737]: W1212 18:39:36.308611 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.308639 kubelet[2737]: E1212 18:39:36.308621 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.309472 kubelet[2737]: E1212 18:39:36.309381 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.309472 kubelet[2737]: W1212 18:39:36.309416 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.309472 kubelet[2737]: E1212 18:39:36.309428 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.309937 kubelet[2737]: E1212 18:39:36.309897 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.309937 kubelet[2737]: W1212 18:39:36.309913 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.309937 kubelet[2737]: E1212 18:39:36.309923 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.310427 kubelet[2737]: E1212 18:39:36.310272 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.310427 kubelet[2737]: W1212 18:39:36.310283 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.310427 kubelet[2737]: E1212 18:39:36.310292 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:36.310498 kubelet[2737]: E1212 18:39:36.310460 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:36.310498 kubelet[2737]: W1212 18:39:36.310469 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:36.310498 kubelet[2737]: E1212 18:39:36.310480 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.119233 kubelet[2737]: E1212 18:39:37.118749 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:39:37.181727 containerd[1549]: time="2025-12-12T18:39:37.181677961Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:37.182926 containerd[1549]: time="2025-12-12T18:39:37.182825537Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=4446754" Dec 12 18:39:37.183878 containerd[1549]: time="2025-12-12T18:39:37.183851045Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:37.185565 containerd[1549]: time="2025-12-12T18:39:37.185530496Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:37.186161 containerd[1549]: time="2025-12-12T18:39:37.186132833Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 1.568161123s" Dec 12 18:39:37.186283 containerd[1549]: time="2025-12-12T18:39:37.186265490Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Dec 12 18:39:37.192711 containerd[1549]: time="2025-12-12T18:39:37.192675688Z" level=info msg="CreateContainer within sandbox \"508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Dec 12 18:39:37.202125 containerd[1549]: time="2025-12-12T18:39:37.201268948Z" level=info msg="Container 99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:37.217880 containerd[1549]: time="2025-12-12T18:39:37.217832714Z" level=info msg="CreateContainer within sandbox \"508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94\"" Dec 12 18:39:37.218626 containerd[1549]: time="2025-12-12T18:39:37.218592645Z" level=info msg="StartContainer for \"99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94\"" Dec 12 18:39:37.220993 containerd[1549]: time="2025-12-12T18:39:37.220959010Z" level=info msg="connecting to shim 99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94" address="unix:///run/containerd/s/343d31595ab3d1ab90348b49fd8f71e383b8f1e0e2ab4a166558feae6c86541d" protocol=ttrpc version=3 Dec 12 18:39:37.239549 kubelet[2737]: I1212 18:39:37.239102 2737 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 18:39:37.259340 systemd[1]: Started cri-containerd-99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94.scope - libcontainer container 99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94. Dec 12 18:39:37.295702 kubelet[2737]: E1212 18:39:37.295643 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.297260 kubelet[2737]: W1212 18:39:37.295672 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.297260 kubelet[2737]: E1212 18:39:37.296288 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.297260 kubelet[2737]: E1212 18:39:37.296606 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.297260 kubelet[2737]: W1212 18:39:37.296618 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.297260 kubelet[2737]: E1212 18:39:37.296649 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.298169 kubelet[2737]: E1212 18:39:37.297564 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.298169 kubelet[2737]: W1212 18:39:37.297576 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.298169 kubelet[2737]: E1212 18:39:37.297616 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.298169 kubelet[2737]: E1212 18:39:37.297855 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.298169 kubelet[2737]: W1212 18:39:37.297876 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.298169 kubelet[2737]: E1212 18:39:37.297889 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.298932 kubelet[2737]: E1212 18:39:37.298352 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.298932 kubelet[2737]: W1212 18:39:37.298367 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.298932 kubelet[2737]: E1212 18:39:37.298404 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.298932 kubelet[2737]: E1212 18:39:37.298778 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.298932 kubelet[2737]: W1212 18:39:37.298787 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.298932 kubelet[2737]: E1212 18:39:37.298822 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.299380 kubelet[2737]: E1212 18:39:37.299334 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.299380 kubelet[2737]: W1212 18:39:37.299346 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.299380 kubelet[2737]: E1212 18:39:37.299359 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.299736 kubelet[2737]: E1212 18:39:37.299697 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.299736 kubelet[2737]: W1212 18:39:37.299708 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.299925 kubelet[2737]: E1212 18:39:37.299720 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.300159 kubelet[2737]: E1212 18:39:37.300125 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.300159 kubelet[2737]: W1212 18:39:37.300136 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.300382 kubelet[2737]: E1212 18:39:37.300278 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.300530 kubelet[2737]: E1212 18:39:37.300493 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.300530 kubelet[2737]: W1212 18:39:37.300504 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.300675 kubelet[2737]: E1212 18:39:37.300514 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.300782 kubelet[2737]: E1212 18:39:37.300772 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.300937 kubelet[2737]: W1212 18:39:37.300840 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.300937 kubelet[2737]: E1212 18:39:37.300852 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.302318 kubelet[2737]: E1212 18:39:37.302290 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.302371 kubelet[2737]: W1212 18:39:37.302307 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.302371 kubelet[2737]: E1212 18:39:37.302338 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.302902 kubelet[2737]: E1212 18:39:37.302537 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.302902 kubelet[2737]: W1212 18:39:37.302545 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.302902 kubelet[2737]: E1212 18:39:37.302553 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.302902 kubelet[2737]: E1212 18:39:37.302786 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.302902 kubelet[2737]: W1212 18:39:37.302808 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.302902 kubelet[2737]: E1212 18:39:37.302817 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.303061 kubelet[2737]: E1212 18:39:37.302959 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.303061 kubelet[2737]: W1212 18:39:37.302966 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.303061 kubelet[2737]: E1212 18:39:37.302974 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.315485 kubelet[2737]: E1212 18:39:37.315454 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.315485 kubelet[2737]: W1212 18:39:37.315472 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.316451 kubelet[2737]: E1212 18:39:37.315486 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.316451 kubelet[2737]: E1212 18:39:37.315707 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.316451 kubelet[2737]: W1212 18:39:37.315716 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.316451 kubelet[2737]: E1212 18:39:37.315724 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.316451 kubelet[2737]: E1212 18:39:37.316091 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.316451 kubelet[2737]: W1212 18:39:37.316264 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.316451 kubelet[2737]: E1212 18:39:37.316280 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.316769 kubelet[2737]: E1212 18:39:37.316715 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.316769 kubelet[2737]: W1212 18:39:37.316728 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.316769 kubelet[2737]: E1212 18:39:37.316738 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.317462 kubelet[2737]: E1212 18:39:37.317429 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.317462 kubelet[2737]: W1212 18:39:37.317442 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.317462 kubelet[2737]: E1212 18:39:37.317452 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.317913 kubelet[2737]: E1212 18:39:37.317872 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.317913 kubelet[2737]: W1212 18:39:37.317886 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.318128 kubelet[2737]: E1212 18:39:37.317896 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.318363 kubelet[2737]: E1212 18:39:37.318350 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.318415 kubelet[2737]: W1212 18:39:37.318365 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.318415 kubelet[2737]: E1212 18:39:37.318374 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.318600 kubelet[2737]: E1212 18:39:37.318557 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.318600 kubelet[2737]: W1212 18:39:37.318593 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.318679 kubelet[2737]: E1212 18:39:37.318603 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.319783 kubelet[2737]: E1212 18:39:37.319647 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.319783 kubelet[2737]: W1212 18:39:37.319708 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.319783 kubelet[2737]: E1212 18:39:37.319725 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.320412 kubelet[2737]: E1212 18:39:37.320375 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.320412 kubelet[2737]: W1212 18:39:37.320408 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.320623 kubelet[2737]: E1212 18:39:37.320421 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.321304 kubelet[2737]: E1212 18:39:37.321290 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.321479 kubelet[2737]: W1212 18:39:37.321364 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.321479 kubelet[2737]: E1212 18:39:37.321382 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.321592 kubelet[2737]: E1212 18:39:37.321573 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.321592 kubelet[2737]: W1212 18:39:37.321587 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.321889 kubelet[2737]: E1212 18:39:37.321597 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.321889 kubelet[2737]: E1212 18:39:37.321776 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.321889 kubelet[2737]: W1212 18:39:37.321784 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.321889 kubelet[2737]: E1212 18:39:37.321792 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.322340 kubelet[2737]: E1212 18:39:37.322318 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.322340 kubelet[2737]: W1212 18:39:37.322330 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.322340 kubelet[2737]: E1212 18:39:37.322340 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.323006 kubelet[2737]: E1212 18:39:37.322507 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.323006 kubelet[2737]: W1212 18:39:37.322516 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.323006 kubelet[2737]: E1212 18:39:37.322524 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.323142 kubelet[2737]: E1212 18:39:37.323038 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.323142 kubelet[2737]: W1212 18:39:37.323047 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.323142 kubelet[2737]: E1212 18:39:37.323056 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.323945 kubelet[2737]: E1212 18:39:37.323924 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.323945 kubelet[2737]: W1212 18:39:37.323939 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.324061 kubelet[2737]: E1212 18:39:37.323952 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.324400 kubelet[2737]: E1212 18:39:37.324386 2737 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 12 18:39:37.324527 kubelet[2737]: W1212 18:39:37.324511 2737 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 12 18:39:37.324662 kubelet[2737]: E1212 18:39:37.324630 2737 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 12 18:39:37.336157 containerd[1549]: time="2025-12-12T18:39:37.336100065Z" level=info msg="StartContainer for \"99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94\" returns successfully" Dec 12 18:39:37.355205 systemd[1]: cri-containerd-99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94.scope: Deactivated successfully. Dec 12 18:39:37.389674 containerd[1549]: time="2025-12-12T18:39:37.389482146Z" level=info msg="received container exit event container_id:\"99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94\" id:\"99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94\" pid:3441 exited_at:{seconds:1765564777 nanos:359096281}" Dec 12 18:39:37.432909 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-99365107ea64c002c1f482ff51c6c53a26eee14fe0b902b5f3d6e1fee039be94-rootfs.mount: Deactivated successfully. Dec 12 18:39:38.244873 containerd[1549]: time="2025-12-12T18:39:38.244834326Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Dec 12 18:39:39.119868 kubelet[2737]: E1212 18:39:39.119761 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:39:40.957878 containerd[1549]: time="2025-12-12T18:39:40.957390884Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:40.958295 containerd[1549]: time="2025-12-12T18:39:40.958269929Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70446859" Dec 12 18:39:40.958898 containerd[1549]: time="2025-12-12T18:39:40.958848652Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:40.960534 containerd[1549]: time="2025-12-12T18:39:40.960499521Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:40.961037 containerd[1549]: time="2025-12-12T18:39:40.960864073Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 2.715992737s" Dec 12 18:39:40.961037 containerd[1549]: time="2025-12-12T18:39:40.960886655Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Dec 12 18:39:40.970558 containerd[1549]: time="2025-12-12T18:39:40.970533865Z" level=info msg="CreateContainer within sandbox \"508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Dec 12 18:39:40.985089 containerd[1549]: time="2025-12-12T18:39:40.984913900Z" level=info msg="Container 924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:40.994059 containerd[1549]: time="2025-12-12T18:39:40.994017883Z" level=info msg="CreateContainer within sandbox \"508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8\"" Dec 12 18:39:40.995549 containerd[1549]: time="2025-12-12T18:39:40.994558886Z" level=info msg="StartContainer for \"924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8\"" Dec 12 18:39:40.996933 containerd[1549]: time="2025-12-12T18:39:40.996904965Z" level=info msg="connecting to shim 924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8" address="unix:///run/containerd/s/343d31595ab3d1ab90348b49fd8f71e383b8f1e0e2ab4a166558feae6c86541d" protocol=ttrpc version=3 Dec 12 18:39:41.021332 systemd[1]: Started cri-containerd-924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8.scope - libcontainer container 924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8. Dec 12 18:39:41.076164 containerd[1549]: time="2025-12-12T18:39:41.076131245Z" level=info msg="StartContainer for \"924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8\" returns successfully" Dec 12 18:39:41.119550 kubelet[2737]: E1212 18:39:41.119510 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:39:41.445883 systemd[1]: cri-containerd-924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8.scope: Deactivated successfully. Dec 12 18:39:41.446219 systemd[1]: cri-containerd-924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8.scope: Consumed 351ms CPU time, 158.9M memory peak, 12K read from disk, 171.3M written to disk. Dec 12 18:39:41.517203 containerd[1549]: time="2025-12-12T18:39:41.517159353Z" level=info msg="received container exit event container_id:\"924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8\" id:\"924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8\" pid:3531 exited_at:{seconds:1765564781 nanos:506719098}" Dec 12 18:39:41.517753 kubelet[2737]: I1212 18:39:41.517733 2737 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Dec 12 18:39:41.590391 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-924364c011c749f9709d296a0714081ff356db5525d7e80abfd96ef2617575d8-rootfs.mount: Deactivated successfully. Dec 12 18:39:41.598990 systemd[1]: Created slice kubepods-burstable-pod5da10053_0265_4285_92bb_b399b80bb162.slice - libcontainer container kubepods-burstable-pod5da10053_0265_4285_92bb_b399b80bb162.slice. Dec 12 18:39:41.625240 systemd[1]: Created slice kubepods-besteffort-pod53428c7e_77fb_479f_978b_3877d65c8e58.slice - libcontainer container kubepods-besteffort-pod53428c7e_77fb_479f_978b_3877d65c8e58.slice. Dec 12 18:39:41.633783 systemd[1]: Created slice kubepods-besteffort-podae09f5bd_f7e2_460b_b94b_eb370fa59677.slice - libcontainer container kubepods-besteffort-podae09f5bd_f7e2_460b_b94b_eb370fa59677.slice. Dec 12 18:39:41.640420 systemd[1]: Created slice kubepods-burstable-pod3431bd32_21c9_4310_bd4d_682f372a9f24.slice - libcontainer container kubepods-burstable-pod3431bd32_21c9_4310_bd4d_682f372a9f24.slice. Dec 12 18:39:41.646598 systemd[1]: Created slice kubepods-besteffort-pod3a064600_3d57_4e0d_80c2_a8041300095d.slice - libcontainer container kubepods-besteffort-pod3a064600_3d57_4e0d_80c2_a8041300095d.slice. Dec 12 18:39:41.651362 kubelet[2737]: I1212 18:39:41.651139 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5da10053-0265-4285-92bb-b399b80bb162-config-volume\") pod \"coredns-674b8bbfcf-24wrg\" (UID: \"5da10053-0265-4285-92bb-b399b80bb162\") " pod="kube-system/coredns-674b8bbfcf-24wrg" Dec 12 18:39:41.651362 kubelet[2737]: I1212 18:39:41.651171 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3431bd32-21c9-4310-bd4d-682f372a9f24-config-volume\") pod \"coredns-674b8bbfcf-f26wm\" (UID: \"3431bd32-21c9-4310-bd4d-682f372a9f24\") " pod="kube-system/coredns-674b8bbfcf-f26wm" Dec 12 18:39:41.652069 kubelet[2737]: I1212 18:39:41.651944 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a064600-3d57-4e0d-80c2-a8041300095d-config\") pod \"goldmane-666569f655-92gml\" (UID: \"3a064600-3d57-4e0d-80c2-a8041300095d\") " pod="calico-system/goldmane-666569f655-92gml" Dec 12 18:39:41.652069 kubelet[2737]: I1212 18:39:41.651973 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mwjv\" (UniqueName: \"kubernetes.io/projected/53428c7e-77fb-479f-978b-3877d65c8e58-kube-api-access-7mwjv\") pod \"calico-kube-controllers-7bbfbd7b8d-c5gdp\" (UID: \"53428c7e-77fb-479f-978b-3877d65c8e58\") " pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" Dec 12 18:39:41.652069 kubelet[2737]: I1212 18:39:41.651989 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d5ds\" (UniqueName: \"kubernetes.io/projected/3a064600-3d57-4e0d-80c2-a8041300095d-kube-api-access-5d5ds\") pod \"goldmane-666569f655-92gml\" (UID: \"3a064600-3d57-4e0d-80c2-a8041300095d\") " pod="calico-system/goldmane-666569f655-92gml" Dec 12 18:39:41.653293 systemd[1]: Created slice kubepods-besteffort-pod986a62cf_f52b_49ec_a518_51f0449c9112.slice - libcontainer container kubepods-besteffort-pod986a62cf_f52b_49ec_a518_51f0449c9112.slice. Dec 12 18:39:41.654026 kubelet[2737]: I1212 18:39:41.654002 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/ae09f5bd-f7e2-460b-b94b-eb370fa59677-calico-apiserver-certs\") pod \"calico-apiserver-59789c7c4b-4fs8q\" (UID: \"ae09f5bd-f7e2-460b-b94b-eb370fa59677\") " pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" Dec 12 18:39:41.654120 kubelet[2737]: I1212 18:39:41.654090 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a064600-3d57-4e0d-80c2-a8041300095d-goldmane-ca-bundle\") pod \"goldmane-666569f655-92gml\" (UID: \"3a064600-3d57-4e0d-80c2-a8041300095d\") " pod="calico-system/goldmane-666569f655-92gml" Dec 12 18:39:41.654220 kubelet[2737]: I1212 18:39:41.654175 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-whisker-backend-key-pair\") pod \"whisker-7b4f57f94d-vkhrh\" (UID: \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\") " pod="calico-system/whisker-7b4f57f94d-vkhrh" Dec 12 18:39:41.654572 kubelet[2737]: I1212 18:39:41.654306 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhmrw\" (UniqueName: \"kubernetes.io/projected/3431bd32-21c9-4310-bd4d-682f372a9f24-kube-api-access-nhmrw\") pod \"coredns-674b8bbfcf-f26wm\" (UID: \"3431bd32-21c9-4310-bd4d-682f372a9f24\") " pod="kube-system/coredns-674b8bbfcf-f26wm" Dec 12 18:39:41.654572 kubelet[2737]: I1212 18:39:41.654324 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bblzv\" (UniqueName: \"kubernetes.io/projected/5da10053-0265-4285-92bb-b399b80bb162-kube-api-access-bblzv\") pod \"coredns-674b8bbfcf-24wrg\" (UID: \"5da10053-0265-4285-92bb-b399b80bb162\") " pod="kube-system/coredns-674b8bbfcf-24wrg" Dec 12 18:39:41.654572 kubelet[2737]: I1212 18:39:41.654335 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-whisker-ca-bundle\") pod \"whisker-7b4f57f94d-vkhrh\" (UID: \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\") " pod="calico-system/whisker-7b4f57f94d-vkhrh" Dec 12 18:39:41.654572 kubelet[2737]: I1212 18:39:41.654346 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvxl6\" (UniqueName: \"kubernetes.io/projected/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-kube-api-access-kvxl6\") pod \"whisker-7b4f57f94d-vkhrh\" (UID: \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\") " pod="calico-system/whisker-7b4f57f94d-vkhrh" Dec 12 18:39:41.654572 kubelet[2737]: I1212 18:39:41.654384 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lpjp\" (UniqueName: \"kubernetes.io/projected/986a62cf-f52b-49ec-a518-51f0449c9112-kube-api-access-9lpjp\") pod \"calico-apiserver-59789c7c4b-wht4c\" (UID: \"986a62cf-f52b-49ec-a518-51f0449c9112\") " pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" Dec 12 18:39:41.654691 kubelet[2737]: I1212 18:39:41.654406 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53428c7e-77fb-479f-978b-3877d65c8e58-tigera-ca-bundle\") pod \"calico-kube-controllers-7bbfbd7b8d-c5gdp\" (UID: \"53428c7e-77fb-479f-978b-3877d65c8e58\") " pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" Dec 12 18:39:41.654691 kubelet[2737]: I1212 18:39:41.654419 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/3a064600-3d57-4e0d-80c2-a8041300095d-goldmane-key-pair\") pod \"goldmane-666569f655-92gml\" (UID: \"3a064600-3d57-4e0d-80c2-a8041300095d\") " pod="calico-system/goldmane-666569f655-92gml" Dec 12 18:39:41.654691 kubelet[2737]: I1212 18:39:41.654430 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfpwb\" (UniqueName: \"kubernetes.io/projected/ae09f5bd-f7e2-460b-b94b-eb370fa59677-kube-api-access-hfpwb\") pod \"calico-apiserver-59789c7c4b-4fs8q\" (UID: \"ae09f5bd-f7e2-460b-b94b-eb370fa59677\") " pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" Dec 12 18:39:41.654691 kubelet[2737]: I1212 18:39:41.654442 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/986a62cf-f52b-49ec-a518-51f0449c9112-calico-apiserver-certs\") pod \"calico-apiserver-59789c7c4b-wht4c\" (UID: \"986a62cf-f52b-49ec-a518-51f0449c9112\") " pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" Dec 12 18:39:41.660727 systemd[1]: Created slice kubepods-besteffort-pod5cbf3f39_fefd_42fd_bb4d_a1e6131af206.slice - libcontainer container kubepods-besteffort-pod5cbf3f39_fefd_42fd_bb4d_a1e6131af206.slice. Dec 12 18:39:41.931112 containerd[1549]: time="2025-12-12T18:39:41.931060181Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7bbfbd7b8d-c5gdp,Uid:53428c7e-77fb-479f-978b-3877d65c8e58,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:41.936209 containerd[1549]: time="2025-12-12T18:39:41.933247042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-24wrg,Uid:5da10053-0265-4285-92bb-b399b80bb162,Namespace:kube-system,Attempt:0,}" Dec 12 18:39:41.940024 containerd[1549]: time="2025-12-12T18:39:41.939976595Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59789c7c4b-4fs8q,Uid:ae09f5bd-f7e2-460b-b94b-eb370fa59677,Namespace:calico-apiserver,Attempt:0,}" Dec 12 18:39:41.948701 containerd[1549]: time="2025-12-12T18:39:41.948483894Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-f26wm,Uid:3431bd32-21c9-4310-bd4d-682f372a9f24,Namespace:kube-system,Attempt:0,}" Dec 12 18:39:41.960038 containerd[1549]: time="2025-12-12T18:39:41.959986168Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59789c7c4b-wht4c,Uid:986a62cf-f52b-49ec-a518-51f0449c9112,Namespace:calico-apiserver,Attempt:0,}" Dec 12 18:39:41.960846 containerd[1549]: time="2025-12-12T18:39:41.960814518Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-92gml,Uid:3a064600-3d57-4e0d-80c2-a8041300095d,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:41.963558 containerd[1549]: time="2025-12-12T18:39:41.963498530Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7b4f57f94d-vkhrh,Uid:5cbf3f39-fefd-42fd-bb4d-a1e6131af206,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:42.199402 containerd[1549]: time="2025-12-12T18:39:42.199234386Z" level=error msg="Failed to destroy network for sandbox \"bd84d423b41fd9581a713536d2d134afbe15e7d842cab98c471ba10dc1f0e71f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.201909 systemd[1]: run-netns-cni\x2d9cf61e1b\x2d0ba1\x2d643e\x2d6caf\x2de516eec484bb.mount: Deactivated successfully. Dec 12 18:39:42.203643 containerd[1549]: time="2025-12-12T18:39:42.203616037Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59789c7c4b-wht4c,Uid:986a62cf-f52b-49ec-a518-51f0449c9112,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bd84d423b41fd9581a713536d2d134afbe15e7d842cab98c471ba10dc1f0e71f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.204837 kubelet[2737]: E1212 18:39:42.204621 2737 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bd84d423b41fd9581a713536d2d134afbe15e7d842cab98c471ba10dc1f0e71f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.204837 kubelet[2737]: E1212 18:39:42.204702 2737 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bd84d423b41fd9581a713536d2d134afbe15e7d842cab98c471ba10dc1f0e71f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" Dec 12 18:39:42.204837 kubelet[2737]: E1212 18:39:42.204722 2737 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bd84d423b41fd9581a713536d2d134afbe15e7d842cab98c471ba10dc1f0e71f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" Dec 12 18:39:42.205091 kubelet[2737]: E1212 18:39:42.204807 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-59789c7c4b-wht4c_calico-apiserver(986a62cf-f52b-49ec-a518-51f0449c9112)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-59789c7c4b-wht4c_calico-apiserver(986a62cf-f52b-49ec-a518-51f0449c9112)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bd84d423b41fd9581a713536d2d134afbe15e7d842cab98c471ba10dc1f0e71f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:39:42.215446 containerd[1549]: time="2025-12-12T18:39:42.215400852Z" level=error msg="Failed to destroy network for sandbox \"3055b37d62fbbad523f6b5271e939db5f26ed92d5c255582f8ebb29d015462ae\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.217058 systemd[1]: run-netns-cni\x2dc8465ccc\x2dac44\x2d8f0d\x2da815\x2d818c54daf99c.mount: Deactivated successfully. Dec 12 18:39:42.220419 containerd[1549]: time="2025-12-12T18:39:42.220309848Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7bbfbd7b8d-c5gdp,Uid:53428c7e-77fb-479f-978b-3877d65c8e58,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3055b37d62fbbad523f6b5271e939db5f26ed92d5c255582f8ebb29d015462ae\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.220972 kubelet[2737]: E1212 18:39:42.220482 2737 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3055b37d62fbbad523f6b5271e939db5f26ed92d5c255582f8ebb29d015462ae\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.220972 kubelet[2737]: E1212 18:39:42.220533 2737 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3055b37d62fbbad523f6b5271e939db5f26ed92d5c255582f8ebb29d015462ae\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" Dec 12 18:39:42.220972 kubelet[2737]: E1212 18:39:42.220551 2737 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3055b37d62fbbad523f6b5271e939db5f26ed92d5c255582f8ebb29d015462ae\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" Dec 12 18:39:42.221294 kubelet[2737]: E1212 18:39:42.220599 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7bbfbd7b8d-c5gdp_calico-system(53428c7e-77fb-479f-978b-3877d65c8e58)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7bbfbd7b8d-c5gdp_calico-system(53428c7e-77fb-479f-978b-3877d65c8e58)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3055b37d62fbbad523f6b5271e939db5f26ed92d5c255582f8ebb29d015462ae\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:39:42.222978 containerd[1549]: time="2025-12-12T18:39:42.222323627Z" level=error msg="Failed to destroy network for sandbox \"ac80a26a8ec606cdb27d338cc3710ab2198fdb9baa93f5df84b2d457006e7935\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.222978 containerd[1549]: time="2025-12-12T18:39:42.222444183Z" level=error msg="Failed to destroy network for sandbox \"78c9364a9663c17bb77778720676573a8569e2e62d061f9d98381361454f98bc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.226077 containerd[1549]: time="2025-12-12T18:39:42.226043490Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-24wrg,Uid:5da10053-0265-4285-92bb-b399b80bb162,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"78c9364a9663c17bb77778720676573a8569e2e62d061f9d98381361454f98bc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.227021 containerd[1549]: time="2025-12-12T18:39:42.226995160Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7b4f57f94d-vkhrh,Uid:5cbf3f39-fefd-42fd-bb4d-a1e6131af206,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac80a26a8ec606cdb27d338cc3710ab2198fdb9baa93f5df84b2d457006e7935\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.227164 kubelet[2737]: E1212 18:39:42.227137 2737 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac80a26a8ec606cdb27d338cc3710ab2198fdb9baa93f5df84b2d457006e7935\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.227164 kubelet[2737]: E1212 18:39:42.227173 2737 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac80a26a8ec606cdb27d338cc3710ab2198fdb9baa93f5df84b2d457006e7935\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7b4f57f94d-vkhrh" Dec 12 18:39:42.227422 kubelet[2737]: E1212 18:39:42.227137 2737 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"78c9364a9663c17bb77778720676573a8569e2e62d061f9d98381361454f98bc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.227422 kubelet[2737]: E1212 18:39:42.227207 2737 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac80a26a8ec606cdb27d338cc3710ab2198fdb9baa93f5df84b2d457006e7935\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7b4f57f94d-vkhrh" Dec 12 18:39:42.227422 kubelet[2737]: E1212 18:39:42.227233 2737 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"78c9364a9663c17bb77778720676573a8569e2e62d061f9d98381361454f98bc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-24wrg" Dec 12 18:39:42.227422 kubelet[2737]: E1212 18:39:42.227255 2737 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"78c9364a9663c17bb77778720676573a8569e2e62d061f9d98381361454f98bc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-24wrg" Dec 12 18:39:42.227954 kubelet[2737]: E1212 18:39:42.227252 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-7b4f57f94d-vkhrh_calico-system(5cbf3f39-fefd-42fd-bb4d-a1e6131af206)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-7b4f57f94d-vkhrh_calico-system(5cbf3f39-fefd-42fd-bb4d-a1e6131af206)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ac80a26a8ec606cdb27d338cc3710ab2198fdb9baa93f5df84b2d457006e7935\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7b4f57f94d-vkhrh" podUID="5cbf3f39-fefd-42fd-bb4d-a1e6131af206" Dec 12 18:39:42.227954 kubelet[2737]: E1212 18:39:42.227301 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-24wrg_kube-system(5da10053-0265-4285-92bb-b399b80bb162)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-24wrg_kube-system(5da10053-0265-4285-92bb-b399b80bb162)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"78c9364a9663c17bb77778720676573a8569e2e62d061f9d98381361454f98bc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-24wrg" podUID="5da10053-0265-4285-92bb-b399b80bb162" Dec 12 18:39:42.229709 containerd[1549]: time="2025-12-12T18:39:42.229643095Z" level=error msg="Failed to destroy network for sandbox \"70d1c7e42180564834808c45ebf9ea892f2a718359420d5de17cccdf5a6dc70c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.229763 containerd[1549]: time="2025-12-12T18:39:42.229747060Z" level=error msg="Failed to destroy network for sandbox \"188ff3a581b05d3edbb70b8b1c350ca431a809cf58dd2355022e0d4037816bd1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.231226 containerd[1549]: time="2025-12-12T18:39:42.230669657Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-92gml,Uid:3a064600-3d57-4e0d-80c2-a8041300095d,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"188ff3a581b05d3edbb70b8b1c350ca431a809cf58dd2355022e0d4037816bd1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.231290 kubelet[2737]: E1212 18:39:42.230850 2737 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"188ff3a581b05d3edbb70b8b1c350ca431a809cf58dd2355022e0d4037816bd1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.231290 kubelet[2737]: E1212 18:39:42.230875 2737 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"188ff3a581b05d3edbb70b8b1c350ca431a809cf58dd2355022e0d4037816bd1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-92gml" Dec 12 18:39:42.231290 kubelet[2737]: E1212 18:39:42.230885 2737 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"188ff3a581b05d3edbb70b8b1c350ca431a809cf58dd2355022e0d4037816bd1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-92gml" Dec 12 18:39:42.231356 kubelet[2737]: E1212 18:39:42.230909 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-92gml_calico-system(3a064600-3d57-4e0d-80c2-a8041300095d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-92gml_calico-system(3a064600-3d57-4e0d-80c2-a8041300095d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"188ff3a581b05d3edbb70b8b1c350ca431a809cf58dd2355022e0d4037816bd1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:39:42.231503 containerd[1549]: time="2025-12-12T18:39:42.231474423Z" level=error msg="Failed to destroy network for sandbox \"a589f1fe210f831c0bcf17c54049ccbe206a947772b002df3b6fb9e104fdb1d8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.231644 containerd[1549]: time="2025-12-12T18:39:42.231624122Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-f26wm,Uid:3431bd32-21c9-4310-bd4d-682f372a9f24,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"70d1c7e42180564834808c45ebf9ea892f2a718359420d5de17cccdf5a6dc70c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.232122 kubelet[2737]: E1212 18:39:42.231869 2737 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70d1c7e42180564834808c45ebf9ea892f2a718359420d5de17cccdf5a6dc70c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.232635 kubelet[2737]: E1212 18:39:42.231893 2737 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70d1c7e42180564834808c45ebf9ea892f2a718359420d5de17cccdf5a6dc70c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-f26wm" Dec 12 18:39:42.232635 kubelet[2737]: E1212 18:39:42.232624 2737 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70d1c7e42180564834808c45ebf9ea892f2a718359420d5de17cccdf5a6dc70c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-f26wm" Dec 12 18:39:42.232729 kubelet[2737]: E1212 18:39:42.232656 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-f26wm_kube-system(3431bd32-21c9-4310-bd4d-682f372a9f24)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-f26wm_kube-system(3431bd32-21c9-4310-bd4d-682f372a9f24)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"70d1c7e42180564834808c45ebf9ea892f2a718359420d5de17cccdf5a6dc70c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-f26wm" podUID="3431bd32-21c9-4310-bd4d-682f372a9f24" Dec 12 18:39:42.233295 containerd[1549]: time="2025-12-12T18:39:42.233263842Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59789c7c4b-4fs8q,Uid:ae09f5bd-f7e2-460b-b94b-eb370fa59677,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a589f1fe210f831c0bcf17c54049ccbe206a947772b002df3b6fb9e104fdb1d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.233434 kubelet[2737]: E1212 18:39:42.233412 2737 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a589f1fe210f831c0bcf17c54049ccbe206a947772b002df3b6fb9e104fdb1d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:42.233470 kubelet[2737]: E1212 18:39:42.233447 2737 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a589f1fe210f831c0bcf17c54049ccbe206a947772b002df3b6fb9e104fdb1d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" Dec 12 18:39:42.233470 kubelet[2737]: E1212 18:39:42.233461 2737 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a589f1fe210f831c0bcf17c54049ccbe206a947772b002df3b6fb9e104fdb1d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" Dec 12 18:39:42.233509 kubelet[2737]: E1212 18:39:42.233485 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-59789c7c4b-4fs8q_calico-apiserver(ae09f5bd-f7e2-460b-b94b-eb370fa59677)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-59789c7c4b-4fs8q_calico-apiserver(ae09f5bd-f7e2-460b-b94b-eb370fa59677)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a589f1fe210f831c0bcf17c54049ccbe206a947772b002df3b6fb9e104fdb1d8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:39:42.277706 containerd[1549]: time="2025-12-12T18:39:42.277679154Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Dec 12 18:39:42.983590 systemd[1]: run-netns-cni\x2d694a9b99\x2dbd04\x2de76a\x2d1495\x2d1afa9cfb060e.mount: Deactivated successfully. Dec 12 18:39:42.983681 systemd[1]: run-netns-cni\x2d7084757c\x2d512d\x2d35b9\x2dcada\x2dbbd9bec77a0b.mount: Deactivated successfully. Dec 12 18:39:42.983733 systemd[1]: run-netns-cni\x2d1bce0577\x2dd3d1\x2d06f4\x2d0fc3\x2d3894acd7559e.mount: Deactivated successfully. Dec 12 18:39:42.983806 systemd[1]: run-netns-cni\x2d8c2cd8a4\x2d9d17\x2dfda2\x2d71cd\x2d35db0a204c89.mount: Deactivated successfully. Dec 12 18:39:42.983859 systemd[1]: run-netns-cni\x2d485d0b1a\x2dd199\x2da331\x2d4f5a\x2d0f63e7942f46.mount: Deactivated successfully. Dec 12 18:39:43.127598 systemd[1]: Created slice kubepods-besteffort-pod583f364c_28f5_44f4_853f_72900703e8a2.slice - libcontainer container kubepods-besteffort-pod583f364c_28f5_44f4_853f_72900703e8a2.slice. Dec 12 18:39:43.131388 containerd[1549]: time="2025-12-12T18:39:43.131338810Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2msts,Uid:583f364c-28f5-44f4-853f-72900703e8a2,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:43.196214 containerd[1549]: time="2025-12-12T18:39:43.194058844Z" level=error msg="Failed to destroy network for sandbox \"982161a1f98ca2e23d9b61ca6e2483da4263ad334d073a4ea7d62ef0fc034cd0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:43.196510 systemd[1]: run-netns-cni\x2d40f6fedf\x2dee77\x2d6d05\x2d64fe\x2dbedf69623140.mount: Deactivated successfully. Dec 12 18:39:43.199800 containerd[1549]: time="2025-12-12T18:39:43.199746469Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2msts,Uid:583f364c-28f5-44f4-853f-72900703e8a2,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"982161a1f98ca2e23d9b61ca6e2483da4263ad334d073a4ea7d62ef0fc034cd0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:43.200246 kubelet[2737]: E1212 18:39:43.200036 2737 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"982161a1f98ca2e23d9b61ca6e2483da4263ad334d073a4ea7d62ef0fc034cd0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 12 18:39:43.200246 kubelet[2737]: E1212 18:39:43.200103 2737 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"982161a1f98ca2e23d9b61ca6e2483da4263ad334d073a4ea7d62ef0fc034cd0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-2msts" Dec 12 18:39:43.200246 kubelet[2737]: E1212 18:39:43.200135 2737 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"982161a1f98ca2e23d9b61ca6e2483da4263ad334d073a4ea7d62ef0fc034cd0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-2msts" Dec 12 18:39:43.200527 kubelet[2737]: E1212 18:39:43.200462 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"982161a1f98ca2e23d9b61ca6e2483da4263ad334d073a4ea7d62ef0fc034cd0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:39:47.730717 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2890358802.mount: Deactivated successfully. Dec 12 18:39:47.878072 containerd[1549]: time="2025-12-12T18:39:47.877753304Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156883675" Dec 12 18:39:47.949889 containerd[1549]: time="2025-12-12T18:39:47.949337735Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:47.969254 containerd[1549]: time="2025-12-12T18:39:47.969206730Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:48.020685 containerd[1549]: time="2025-12-12T18:39:48.012514324Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 5.734672906s" Dec 12 18:39:48.020685 containerd[1549]: time="2025-12-12T18:39:48.018735925Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Dec 12 18:39:48.020685 containerd[1549]: time="2025-12-12T18:39:48.012021120Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:39:48.082281 containerd[1549]: time="2025-12-12T18:39:48.082241047Z" level=info msg="CreateContainer within sandbox \"508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Dec 12 18:39:48.127058 containerd[1549]: time="2025-12-12T18:39:48.127023178Z" level=info msg="Container 69db6f56a8e0f3a95e5951c3fd8b1e3ec7f0ba2d5052a39eb4709bd2c9fc0f10: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:48.128181 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1891216086.mount: Deactivated successfully. Dec 12 18:39:48.166911 containerd[1549]: time="2025-12-12T18:39:48.166860299Z" level=info msg="CreateContainer within sandbox \"508ab381af5df55cc30dafd21e3b069d332e03527ff8529cfcc8efe9005ca277\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"69db6f56a8e0f3a95e5951c3fd8b1e3ec7f0ba2d5052a39eb4709bd2c9fc0f10\"" Dec 12 18:39:48.169794 containerd[1549]: time="2025-12-12T18:39:48.169751113Z" level=info msg="StartContainer for \"69db6f56a8e0f3a95e5951c3fd8b1e3ec7f0ba2d5052a39eb4709bd2c9fc0f10\"" Dec 12 18:39:48.175478 containerd[1549]: time="2025-12-12T18:39:48.175450705Z" level=info msg="connecting to shim 69db6f56a8e0f3a95e5951c3fd8b1e3ec7f0ba2d5052a39eb4709bd2c9fc0f10" address="unix:///run/containerd/s/343d31595ab3d1ab90348b49fd8f71e383b8f1e0e2ab4a166558feae6c86541d" protocol=ttrpc version=3 Dec 12 18:39:48.256465 systemd[1]: Started cri-containerd-69db6f56a8e0f3a95e5951c3fd8b1e3ec7f0ba2d5052a39eb4709bd2c9fc0f10.scope - libcontainer container 69db6f56a8e0f3a95e5951c3fd8b1e3ec7f0ba2d5052a39eb4709bd2c9fc0f10. Dec 12 18:39:48.347950 containerd[1549]: time="2025-12-12T18:39:48.345453310Z" level=info msg="StartContainer for \"69db6f56a8e0f3a95e5951c3fd8b1e3ec7f0ba2d5052a39eb4709bd2c9fc0f10\" returns successfully" Dec 12 18:39:48.537570 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Dec 12 18:39:48.539067 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Dec 12 18:39:48.901671 kubelet[2737]: I1212 18:39:48.901615 2737 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-whisker-backend-key-pair\") pod \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\" (UID: \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\") " Dec 12 18:39:48.902675 kubelet[2737]: I1212 18:39:48.901976 2737 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvxl6\" (UniqueName: \"kubernetes.io/projected/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-kube-api-access-kvxl6\") pod \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\" (UID: \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\") " Dec 12 18:39:48.902675 kubelet[2737]: I1212 18:39:48.902011 2737 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-whisker-ca-bundle\") pod \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\" (UID: \"5cbf3f39-fefd-42fd-bb4d-a1e6131af206\") " Dec 12 18:39:48.915137 kubelet[2737]: I1212 18:39:48.914761 2737 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "5cbf3f39-fefd-42fd-bb4d-a1e6131af206" (UID: "5cbf3f39-fefd-42fd-bb4d-a1e6131af206"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 12 18:39:48.925408 systemd[1]: var-lib-kubelet-pods-5cbf3f39\x2dfefd\x2d42fd\x2dbb4d\x2da1e6131af206-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkvxl6.mount: Deactivated successfully. Dec 12 18:39:48.927462 kubelet[2737]: I1212 18:39:48.924704 2737 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "5cbf3f39-fefd-42fd-bb4d-a1e6131af206" (UID: "5cbf3f39-fefd-42fd-bb4d-a1e6131af206"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 12 18:39:48.927462 kubelet[2737]: I1212 18:39:48.926355 2737 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-kube-api-access-kvxl6" (OuterVolumeSpecName: "kube-api-access-kvxl6") pod "5cbf3f39-fefd-42fd-bb4d-a1e6131af206" (UID: "5cbf3f39-fefd-42fd-bb4d-a1e6131af206"). InnerVolumeSpecName "kube-api-access-kvxl6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 12 18:39:48.928834 systemd[1]: var-lib-kubelet-pods-5cbf3f39\x2dfefd\x2d42fd\x2dbb4d\x2da1e6131af206-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Dec 12 18:39:49.003178 kubelet[2737]: I1212 18:39:49.003127 2737 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kvxl6\" (UniqueName: \"kubernetes.io/projected/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-kube-api-access-kvxl6\") on node \"ci-4459-2-2-7-adcd881523\" DevicePath \"\"" Dec 12 18:39:49.003178 kubelet[2737]: I1212 18:39:49.003157 2737 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-whisker-ca-bundle\") on node \"ci-4459-2-2-7-adcd881523\" DevicePath \"\"" Dec 12 18:39:49.003178 kubelet[2737]: I1212 18:39:49.003167 2737 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5cbf3f39-fefd-42fd-bb4d-a1e6131af206-whisker-backend-key-pair\") on node \"ci-4459-2-2-7-adcd881523\" DevicePath \"\"" Dec 12 18:39:49.124393 systemd[1]: Removed slice kubepods-besteffort-pod5cbf3f39_fefd_42fd_bb4d_a1e6131af206.slice - libcontainer container kubepods-besteffort-pod5cbf3f39_fefd_42fd_bb4d_a1e6131af206.slice. Dec 12 18:39:49.363122 kubelet[2737]: I1212 18:39:49.363061 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-ht5sj" podStartSLOduration=1.792603029 podStartE2EDuration="16.363043157s" podCreationTimestamp="2025-12-12 18:39:33 +0000 UTC" firstStartedPulling="2025-12-12 18:39:33.470309561 +0000 UTC m=+20.449156181" lastFinishedPulling="2025-12-12 18:39:48.040749698 +0000 UTC m=+35.019596309" observedRunningTime="2025-12-12 18:39:49.336156264 +0000 UTC m=+36.315002884" watchObservedRunningTime="2025-12-12 18:39:49.363043157 +0000 UTC m=+36.341889767" Dec 12 18:39:49.445731 systemd[1]: Created slice kubepods-besteffort-pod48676334_6f17_4d47_84e1_7fe85ea764df.slice - libcontainer container kubepods-besteffort-pod48676334_6f17_4d47_84e1_7fe85ea764df.slice. Dec 12 18:39:49.507060 kubelet[2737]: I1212 18:39:49.507002 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86lgn\" (UniqueName: \"kubernetes.io/projected/48676334-6f17-4d47-84e1-7fe85ea764df-kube-api-access-86lgn\") pod \"whisker-774c4fc74b-q78zm\" (UID: \"48676334-6f17-4d47-84e1-7fe85ea764df\") " pod="calico-system/whisker-774c4fc74b-q78zm" Dec 12 18:39:49.507060 kubelet[2737]: I1212 18:39:49.507046 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48676334-6f17-4d47-84e1-7fe85ea764df-whisker-ca-bundle\") pod \"whisker-774c4fc74b-q78zm\" (UID: \"48676334-6f17-4d47-84e1-7fe85ea764df\") " pod="calico-system/whisker-774c4fc74b-q78zm" Dec 12 18:39:49.507060 kubelet[2737]: I1212 18:39:49.507063 2737 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/48676334-6f17-4d47-84e1-7fe85ea764df-whisker-backend-key-pair\") pod \"whisker-774c4fc74b-q78zm\" (UID: \"48676334-6f17-4d47-84e1-7fe85ea764df\") " pod="calico-system/whisker-774c4fc74b-q78zm" Dec 12 18:39:49.749051 containerd[1549]: time="2025-12-12T18:39:49.749011899Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-774c4fc74b-q78zm,Uid:48676334-6f17-4d47-84e1-7fe85ea764df,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:50.046642 systemd-networkd[1447]: cali11926dfe6f5: Link UP Dec 12 18:39:50.047235 systemd-networkd[1447]: cali11926dfe6f5: Gained carrier Dec 12 18:39:50.069262 containerd[1549]: 2025-12-12 18:39:49.800 [INFO][3856] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 12 18:39:50.069262 containerd[1549]: 2025-12-12 18:39:49.829 [INFO][3856] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0 whisker-774c4fc74b- calico-system 48676334-6f17-4d47-84e1-7fe85ea764df 874 0 2025-12-12 18:39:49 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:774c4fc74b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4459-2-2-7-adcd881523 whisker-774c4fc74b-q78zm eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali11926dfe6f5 [] [] }} ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Namespace="calico-system" Pod="whisker-774c4fc74b-q78zm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-" Dec 12 18:39:50.069262 containerd[1549]: 2025-12-12 18:39:49.829 [INFO][3856] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Namespace="calico-system" Pod="whisker-774c4fc74b-q78zm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" Dec 12 18:39:50.069262 containerd[1549]: 2025-12-12 18:39:49.978 [INFO][3868] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" HandleID="k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Workload="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:49.981 [INFO][3868] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" HandleID="k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Workload="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000325740), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459-2-2-7-adcd881523", "pod":"whisker-774c4fc74b-q78zm", "timestamp":"2025-12-12 18:39:49.978500474 +0000 UTC"}, Hostname:"ci-4459-2-2-7-adcd881523", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:49.981 [INFO][3868] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:49.981 [INFO][3868] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:49.982 [INFO][3868] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-2-2-7-adcd881523' Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:49.997 [INFO][3868] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:50.010 [INFO][3868] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:50.015 [INFO][3868] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:50.016 [INFO][3868] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.069516 containerd[1549]: 2025-12-12 18:39:50.019 [INFO][3868] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.070397 containerd[1549]: 2025-12-12 18:39:50.019 [INFO][3868] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.070397 containerd[1549]: 2025-12-12 18:39:50.020 [INFO][3868] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9 Dec 12 18:39:50.070397 containerd[1549]: 2025-12-12 18:39:50.025 [INFO][3868] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.070397 containerd[1549]: 2025-12-12 18:39:50.030 [INFO][3868] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.193/26] block=192.168.109.192/26 handle="k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.070397 containerd[1549]: 2025-12-12 18:39:50.030 [INFO][3868] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.193/26] handle="k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:50.070397 containerd[1549]: 2025-12-12 18:39:50.030 [INFO][3868] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 12 18:39:50.070397 containerd[1549]: 2025-12-12 18:39:50.030 [INFO][3868] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.193/26] IPv6=[] ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" HandleID="k8s-pod-network.575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Workload="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" Dec 12 18:39:50.070537 containerd[1549]: 2025-12-12 18:39:50.033 [INFO][3856] cni-plugin/k8s.go 418: Populated endpoint ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Namespace="calico-system" Pod="whisker-774c4fc74b-q78zm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0", GenerateName:"whisker-774c4fc74b-", Namespace:"calico-system", SelfLink:"", UID:"48676334-6f17-4d47-84e1-7fe85ea764df", ResourceVersion:"874", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"774c4fc74b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"", Pod:"whisker-774c4fc74b-q78zm", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.109.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali11926dfe6f5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:50.070537 containerd[1549]: 2025-12-12 18:39:50.033 [INFO][3856] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.193/32] ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Namespace="calico-system" Pod="whisker-774c4fc74b-q78zm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" Dec 12 18:39:50.070621 containerd[1549]: 2025-12-12 18:39:50.033 [INFO][3856] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali11926dfe6f5 ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Namespace="calico-system" Pod="whisker-774c4fc74b-q78zm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" Dec 12 18:39:50.070621 containerd[1549]: 2025-12-12 18:39:50.048 [INFO][3856] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Namespace="calico-system" Pod="whisker-774c4fc74b-q78zm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" Dec 12 18:39:50.070662 containerd[1549]: 2025-12-12 18:39:50.049 [INFO][3856] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Namespace="calico-system" Pod="whisker-774c4fc74b-q78zm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0", GenerateName:"whisker-774c4fc74b-", Namespace:"calico-system", SelfLink:"", UID:"48676334-6f17-4d47-84e1-7fe85ea764df", ResourceVersion:"874", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"774c4fc74b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9", Pod:"whisker-774c4fc74b-q78zm", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.109.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali11926dfe6f5", MAC:"b2:73:56:6e:89:20", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:50.071715 containerd[1549]: 2025-12-12 18:39:50.063 [INFO][3856] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" Namespace="calico-system" Pod="whisker-774c4fc74b-q78zm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-whisker--774c4fc74b--q78zm-eth0" Dec 12 18:39:50.311261 kubelet[2737]: I1212 18:39:50.309475 2737 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 18:39:50.326642 containerd[1549]: time="2025-12-12T18:39:50.326520240Z" level=info msg="connecting to shim 575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9" address="unix:///run/containerd/s/e6a761f302012e970320002e0b4df6de15e22c0b2368ef771b904dbd1c679cf2" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:50.369521 systemd[1]: Started cri-containerd-575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9.scope - libcontainer container 575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9. Dec 12 18:39:50.484087 containerd[1549]: time="2025-12-12T18:39:50.484039755Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-774c4fc74b-q78zm,Uid:48676334-6f17-4d47-84e1-7fe85ea764df,Namespace:calico-system,Attempt:0,} returns sandbox id \"575310ba9a4cf14be41999f3d1fee653fa352b6c08df82713f266ceaecbff8c9\"" Dec 12 18:39:50.517988 containerd[1549]: time="2025-12-12T18:39:50.517944016Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Dec 12 18:39:50.971964 containerd[1549]: time="2025-12-12T18:39:50.971916239Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:39:50.973234 containerd[1549]: time="2025-12-12T18:39:50.973108512Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Dec 12 18:39:50.973234 containerd[1549]: time="2025-12-12T18:39:50.973204842Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Dec 12 18:39:50.973428 kubelet[2737]: E1212 18:39:50.973321 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 12 18:39:50.973428 kubelet[2737]: E1212 18:39:50.973368 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 12 18:39:50.986930 kubelet[2737]: E1212 18:39:50.986862 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:085cdf68872848a99fb533c1a7535d78,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-86lgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-774c4fc74b-q78zm_calico-system(48676334-6f17-4d47-84e1-7fe85ea764df): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Dec 12 18:39:50.990054 containerd[1549]: time="2025-12-12T18:39:50.990022290Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Dec 12 18:39:51.120947 kubelet[2737]: I1212 18:39:51.120904 2737 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbf3f39-fefd-42fd-bb4d-a1e6131af206" path="/var/lib/kubelet/pods/5cbf3f39-fefd-42fd-bb4d-a1e6131af206/volumes" Dec 12 18:39:51.435447 containerd[1549]: time="2025-12-12T18:39:51.435346985Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:39:51.436521 containerd[1549]: time="2025-12-12T18:39:51.436467304Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Dec 12 18:39:51.436521 containerd[1549]: time="2025-12-12T18:39:51.436495277Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Dec 12 18:39:51.436938 kubelet[2737]: E1212 18:39:51.436682 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 12 18:39:51.436938 kubelet[2737]: E1212 18:39:51.436722 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 12 18:39:51.437170 kubelet[2737]: E1212 18:39:51.436850 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86lgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-774c4fc74b-q78zm_calico-system(48676334-6f17-4d47-84e1-7fe85ea764df): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Dec 12 18:39:51.443095 kubelet[2737]: E1212 18:39:51.442808 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:39:51.828353 systemd-networkd[1447]: cali11926dfe6f5: Gained IPv6LL Dec 12 18:39:52.315729 kubelet[2737]: E1212 18:39:52.315679 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:39:54.119334 containerd[1549]: time="2025-12-12T18:39:54.119252097Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-24wrg,Uid:5da10053-0265-4285-92bb-b399b80bb162,Namespace:kube-system,Attempt:0,}" Dec 12 18:39:54.119838 containerd[1549]: time="2025-12-12T18:39:54.119670702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-92gml,Uid:3a064600-3d57-4e0d-80c2-a8041300095d,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:54.279561 systemd-networkd[1447]: calieff6629950c: Link UP Dec 12 18:39:54.280139 systemd-networkd[1447]: calieff6629950c: Gained carrier Dec 12 18:39:54.302145 containerd[1549]: 2025-12-12 18:39:54.165 [INFO][4091] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 12 18:39:54.302145 containerd[1549]: 2025-12-12 18:39:54.182 [INFO][4091] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0 coredns-674b8bbfcf- kube-system 5da10053-0265-4285-92bb-b399b80bb162 802 0 2025-12-12 18:39:19 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4459-2-2-7-adcd881523 coredns-674b8bbfcf-24wrg eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calieff6629950c [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Namespace="kube-system" Pod="coredns-674b8bbfcf-24wrg" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-" Dec 12 18:39:54.302145 containerd[1549]: 2025-12-12 18:39:54.182 [INFO][4091] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Namespace="kube-system" Pod="coredns-674b8bbfcf-24wrg" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" Dec 12 18:39:54.302145 containerd[1549]: 2025-12-12 18:39:54.221 [INFO][4119] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" HandleID="k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Workload="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.222 [INFO][4119] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" HandleID="k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Workload="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002cd5a0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4459-2-2-7-adcd881523", "pod":"coredns-674b8bbfcf-24wrg", "timestamp":"2025-12-12 18:39:54.221870276 +0000 UTC"}, Hostname:"ci-4459-2-2-7-adcd881523", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.222 [INFO][4119] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.222 [INFO][4119] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.222 [INFO][4119] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-2-2-7-adcd881523' Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.231 [INFO][4119] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.241 [INFO][4119] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.250 [INFO][4119] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.252 [INFO][4119] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.302346 containerd[1549]: 2025-12-12 18:39:54.255 [INFO][4119] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.303632 containerd[1549]: 2025-12-12 18:39:54.256 [INFO][4119] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.303632 containerd[1549]: 2025-12-12 18:39:54.257 [INFO][4119] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e Dec 12 18:39:54.303632 containerd[1549]: 2025-12-12 18:39:54.263 [INFO][4119] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.303632 containerd[1549]: 2025-12-12 18:39:54.270 [INFO][4119] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.194/26] block=192.168.109.192/26 handle="k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.303632 containerd[1549]: 2025-12-12 18:39:54.270 [INFO][4119] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.194/26] handle="k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.303632 containerd[1549]: 2025-12-12 18:39:54.270 [INFO][4119] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 12 18:39:54.303632 containerd[1549]: 2025-12-12 18:39:54.270 [INFO][4119] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.194/26] IPv6=[] ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" HandleID="k8s-pod-network.022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Workload="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" Dec 12 18:39:54.303754 containerd[1549]: 2025-12-12 18:39:54.275 [INFO][4091] cni-plugin/k8s.go 418: Populated endpoint ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Namespace="kube-system" Pod="coredns-674b8bbfcf-24wrg" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"5da10053-0265-4285-92bb-b399b80bb162", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"", Pod:"coredns-674b8bbfcf-24wrg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.109.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calieff6629950c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:54.303754 containerd[1549]: 2025-12-12 18:39:54.276 [INFO][4091] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.194/32] ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Namespace="kube-system" Pod="coredns-674b8bbfcf-24wrg" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" Dec 12 18:39:54.303754 containerd[1549]: 2025-12-12 18:39:54.276 [INFO][4091] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieff6629950c ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Namespace="kube-system" Pod="coredns-674b8bbfcf-24wrg" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" Dec 12 18:39:54.303754 containerd[1549]: 2025-12-12 18:39:54.279 [INFO][4091] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Namespace="kube-system" Pod="coredns-674b8bbfcf-24wrg" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" Dec 12 18:39:54.303754 containerd[1549]: 2025-12-12 18:39:54.280 [INFO][4091] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Namespace="kube-system" Pod="coredns-674b8bbfcf-24wrg" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"5da10053-0265-4285-92bb-b399b80bb162", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e", Pod:"coredns-674b8bbfcf-24wrg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.109.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calieff6629950c", MAC:"a2:cc:4f:2d:ae:5b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:54.303754 containerd[1549]: 2025-12-12 18:39:54.299 [INFO][4091] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" Namespace="kube-system" Pod="coredns-674b8bbfcf-24wrg" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--24wrg-eth0" Dec 12 18:39:54.327585 containerd[1549]: time="2025-12-12T18:39:54.327523079Z" level=info msg="connecting to shim 022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e" address="unix:///run/containerd/s/5093a83b82458d4acb6d04fbcff98f95f1fd2f9e0e8c171f7fb7545e6b872baa" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:54.356939 systemd[1]: Started cri-containerd-022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e.scope - libcontainer container 022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e. Dec 12 18:39:54.383033 systemd-networkd[1447]: calibf3cc64bf3c: Link UP Dec 12 18:39:54.384903 systemd-networkd[1447]: calibf3cc64bf3c: Gained carrier Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.168 [INFO][4093] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.183 [INFO][4093] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0 goldmane-666569f655- calico-system 3a064600-3d57-4e0d-80c2-a8041300095d 809 0 2025-12-12 18:39:31 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4459-2-2-7-adcd881523 goldmane-666569f655-92gml eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calibf3cc64bf3c [] [] }} ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Namespace="calico-system" Pod="goldmane-666569f655-92gml" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.183 [INFO][4093] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Namespace="calico-system" Pod="goldmane-666569f655-92gml" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.233 [INFO][4114] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" HandleID="k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Workload="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.234 [INFO][4114] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" HandleID="k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Workload="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f5d0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459-2-2-7-adcd881523", "pod":"goldmane-666569f655-92gml", "timestamp":"2025-12-12 18:39:54.23390368 +0000 UTC"}, Hostname:"ci-4459-2-2-7-adcd881523", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.234 [INFO][4114] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.270 [INFO][4114] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.270 [INFO][4114] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-2-2-7-adcd881523' Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.335 [INFO][4114] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.343 [INFO][4114] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.350 [INFO][4114] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.355 [INFO][4114] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.359 [INFO][4114] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.359 [INFO][4114] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.362 [INFO][4114] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022 Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.368 [INFO][4114] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.376 [INFO][4114] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.195/26] block=192.168.109.192/26 handle="k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.376 [INFO][4114] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.195/26] handle="k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.376 [INFO][4114] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 12 18:39:54.406116 containerd[1549]: 2025-12-12 18:39:54.376 [INFO][4114] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.195/26] IPv6=[] ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" HandleID="k8s-pod-network.4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Workload="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" Dec 12 18:39:54.407766 containerd[1549]: 2025-12-12 18:39:54.380 [INFO][4093] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Namespace="calico-system" Pod="goldmane-666569f655-92gml" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"3a064600-3d57-4e0d-80c2-a8041300095d", ResourceVersion:"809", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"", Pod:"goldmane-666569f655-92gml", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.109.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calibf3cc64bf3c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:54.407766 containerd[1549]: 2025-12-12 18:39:54.380 [INFO][4093] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.195/32] ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Namespace="calico-system" Pod="goldmane-666569f655-92gml" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" Dec 12 18:39:54.407766 containerd[1549]: 2025-12-12 18:39:54.380 [INFO][4093] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calibf3cc64bf3c ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Namespace="calico-system" Pod="goldmane-666569f655-92gml" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" Dec 12 18:39:54.407766 containerd[1549]: 2025-12-12 18:39:54.391 [INFO][4093] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Namespace="calico-system" Pod="goldmane-666569f655-92gml" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" Dec 12 18:39:54.407766 containerd[1549]: 2025-12-12 18:39:54.392 [INFO][4093] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Namespace="calico-system" Pod="goldmane-666569f655-92gml" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"3a064600-3d57-4e0d-80c2-a8041300095d", ResourceVersion:"809", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022", Pod:"goldmane-666569f655-92gml", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.109.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calibf3cc64bf3c", MAC:"9e:9d:64:6d:6e:3a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:54.407766 containerd[1549]: 2025-12-12 18:39:54.402 [INFO][4093] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" Namespace="calico-system" Pod="goldmane-666569f655-92gml" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-goldmane--666569f655--92gml-eth0" Dec 12 18:39:54.435799 containerd[1549]: time="2025-12-12T18:39:54.435390322Z" level=info msg="connecting to shim 4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022" address="unix:///run/containerd/s/abc31b9ce6c4835f5a0dc0d60f2fba3826596a0f26c70422a8fee8536aca24fa" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:54.445456 containerd[1549]: time="2025-12-12T18:39:54.445433797Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-24wrg,Uid:5da10053-0265-4285-92bb-b399b80bb162,Namespace:kube-system,Attempt:0,} returns sandbox id \"022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e\"" Dec 12 18:39:54.454320 containerd[1549]: time="2025-12-12T18:39:54.454293174Z" level=info msg="CreateContainer within sandbox \"022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 12 18:39:54.461432 systemd[1]: Started cri-containerd-4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022.scope - libcontainer container 4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022. Dec 12 18:39:54.476746 containerd[1549]: time="2025-12-12T18:39:54.476721102Z" level=info msg="Container 3105c4c71e2618c4562a1632235e67f17734796d7a0376e79ab492947889fb39: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:54.487096 containerd[1549]: time="2025-12-12T18:39:54.487010888Z" level=info msg="CreateContainer within sandbox \"022ab29b0757e83b583946b5c0ac159d270e0c4009af3a2da7eb40cdfd3d213e\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"3105c4c71e2618c4562a1632235e67f17734796d7a0376e79ab492947889fb39\"" Dec 12 18:39:54.489953 containerd[1549]: time="2025-12-12T18:39:54.489918116Z" level=info msg="StartContainer for \"3105c4c71e2618c4562a1632235e67f17734796d7a0376e79ab492947889fb39\"" Dec 12 18:39:54.490588 containerd[1549]: time="2025-12-12T18:39:54.490558645Z" level=info msg="connecting to shim 3105c4c71e2618c4562a1632235e67f17734796d7a0376e79ab492947889fb39" address="unix:///run/containerd/s/5093a83b82458d4acb6d04fbcff98f95f1fd2f9e0e8c171f7fb7545e6b872baa" protocol=ttrpc version=3 Dec 12 18:39:54.507335 systemd[1]: Started cri-containerd-3105c4c71e2618c4562a1632235e67f17734796d7a0376e79ab492947889fb39.scope - libcontainer container 3105c4c71e2618c4562a1632235e67f17734796d7a0376e79ab492947889fb39. Dec 12 18:39:54.523293 containerd[1549]: time="2025-12-12T18:39:54.523250752Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-92gml,Uid:3a064600-3d57-4e0d-80c2-a8041300095d,Namespace:calico-system,Attempt:0,} returns sandbox id \"4318ad496a572ab99ffdaee07cbe22f024a81bdcaf66da42526b9059e0963022\"" Dec 12 18:39:54.526563 containerd[1549]: time="2025-12-12T18:39:54.526509877Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Dec 12 18:39:54.555029 containerd[1549]: time="2025-12-12T18:39:54.554866903Z" level=info msg="StartContainer for \"3105c4c71e2618c4562a1632235e67f17734796d7a0376e79ab492947889fb39\" returns successfully" Dec 12 18:39:54.965511 containerd[1549]: time="2025-12-12T18:39:54.965414756Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:39:54.967300 containerd[1549]: time="2025-12-12T18:39:54.967164644Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Dec 12 18:39:54.967415 containerd[1549]: time="2025-12-12T18:39:54.967180774Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Dec 12 18:39:54.967819 kubelet[2737]: E1212 18:39:54.967688 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 12 18:39:54.967819 kubelet[2737]: E1212 18:39:54.967739 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 12 18:39:54.968675 kubelet[2737]: E1212 18:39:54.967937 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5d5ds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-92gml_calico-system(3a064600-3d57-4e0d-80c2-a8041300095d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Dec 12 18:39:54.969881 kubelet[2737]: E1212 18:39:54.969838 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:39:55.325391 kubelet[2737]: E1212 18:39:55.324705 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:39:55.411350 systemd-networkd[1447]: calieff6629950c: Gained IPv6LL Dec 12 18:39:56.120886 containerd[1549]: time="2025-12-12T18:39:56.120616317Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59789c7c4b-wht4c,Uid:986a62cf-f52b-49ec-a518-51f0449c9112,Namespace:calico-apiserver,Attempt:0,}" Dec 12 18:39:56.122124 containerd[1549]: time="2025-12-12T18:39:56.122069480Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7bbfbd7b8d-c5gdp,Uid:53428c7e-77fb-479f-978b-3877d65c8e58,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:56.122323 containerd[1549]: time="2025-12-12T18:39:56.122281987Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59789c7c4b-4fs8q,Uid:ae09f5bd-f7e2-460b-b94b-eb370fa59677,Namespace:calico-apiserver,Attempt:0,}" Dec 12 18:39:56.306344 systemd-networkd[1447]: calibf3cc64bf3c: Gained IPv6LL Dec 12 18:39:56.331529 kubelet[2737]: E1212 18:39:56.331492 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:39:56.343940 systemd-networkd[1447]: cali682565a5b18: Link UP Dec 12 18:39:56.345102 systemd-networkd[1447]: cali682565a5b18: Gained carrier Dec 12 18:39:56.353643 kubelet[2737]: I1212 18:39:56.353597 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-24wrg" podStartSLOduration=37.35358253 podStartE2EDuration="37.35358253s" podCreationTimestamp="2025-12-12 18:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:39:55.352150456 +0000 UTC m=+42.330997076" watchObservedRunningTime="2025-12-12 18:39:56.35358253 +0000 UTC m=+43.332429140" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.204 [INFO][4308] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.225 [INFO][4308] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0 calico-kube-controllers-7bbfbd7b8d- calico-system 53428c7e-77fb-479f-978b-3877d65c8e58 807 0 2025-12-12 18:39:33 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:7bbfbd7b8d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4459-2-2-7-adcd881523 calico-kube-controllers-7bbfbd7b8d-c5gdp eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali682565a5b18 [] [] }} ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Namespace="calico-system" Pod="calico-kube-controllers-7bbfbd7b8d-c5gdp" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.225 [INFO][4308] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Namespace="calico-system" Pod="calico-kube-controllers-7bbfbd7b8d-c5gdp" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.294 [INFO][4350] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" HandleID="k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.294 [INFO][4350] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" HandleID="k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f950), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459-2-2-7-adcd881523", "pod":"calico-kube-controllers-7bbfbd7b8d-c5gdp", "timestamp":"2025-12-12 18:39:56.294693103 +0000 UTC"}, Hostname:"ci-4459-2-2-7-adcd881523", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.295 [INFO][4350] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.295 [INFO][4350] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.295 [INFO][4350] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-2-2-7-adcd881523' Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.302 [INFO][4350] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.315 [INFO][4350] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.320 [INFO][4350] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.321 [INFO][4350] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.323 [INFO][4350] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.323 [INFO][4350] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.324 [INFO][4350] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3 Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.330 [INFO][4350] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.337 [INFO][4350] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.196/26] block=192.168.109.192/26 handle="k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.337 [INFO][4350] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.196/26] handle="k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.338 [INFO][4350] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 12 18:39:56.366133 containerd[1549]: 2025-12-12 18:39:56.338 [INFO][4350] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.196/26] IPv6=[] ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" HandleID="k8s-pod-network.390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" Dec 12 18:39:56.367695 containerd[1549]: 2025-12-12 18:39:56.340 [INFO][4308] cni-plugin/k8s.go 418: Populated endpoint ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Namespace="calico-system" Pod="calico-kube-controllers-7bbfbd7b8d-c5gdp" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0", GenerateName:"calico-kube-controllers-7bbfbd7b8d-", Namespace:"calico-system", SelfLink:"", UID:"53428c7e-77fb-479f-978b-3877d65c8e58", ResourceVersion:"807", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7bbfbd7b8d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"", Pod:"calico-kube-controllers-7bbfbd7b8d-c5gdp", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.109.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali682565a5b18", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:56.367695 containerd[1549]: 2025-12-12 18:39:56.340 [INFO][4308] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.196/32] ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Namespace="calico-system" Pod="calico-kube-controllers-7bbfbd7b8d-c5gdp" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" Dec 12 18:39:56.367695 containerd[1549]: 2025-12-12 18:39:56.340 [INFO][4308] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali682565a5b18 ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Namespace="calico-system" Pod="calico-kube-controllers-7bbfbd7b8d-c5gdp" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" Dec 12 18:39:56.367695 containerd[1549]: 2025-12-12 18:39:56.345 [INFO][4308] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Namespace="calico-system" Pod="calico-kube-controllers-7bbfbd7b8d-c5gdp" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" Dec 12 18:39:56.367695 containerd[1549]: 2025-12-12 18:39:56.346 [INFO][4308] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Namespace="calico-system" Pod="calico-kube-controllers-7bbfbd7b8d-c5gdp" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0", GenerateName:"calico-kube-controllers-7bbfbd7b8d-", Namespace:"calico-system", SelfLink:"", UID:"53428c7e-77fb-479f-978b-3877d65c8e58", ResourceVersion:"807", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7bbfbd7b8d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3", Pod:"calico-kube-controllers-7bbfbd7b8d-c5gdp", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.109.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali682565a5b18", MAC:"f6:20:0b:d2:5c:0d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:56.367695 containerd[1549]: 2025-12-12 18:39:56.361 [INFO][4308] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" Namespace="calico-system" Pod="calico-kube-controllers-7bbfbd7b8d-c5gdp" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--kube--controllers--7bbfbd7b8d--c5gdp-eth0" Dec 12 18:39:56.389400 containerd[1549]: time="2025-12-12T18:39:56.388388905Z" level=info msg="connecting to shim 390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3" address="unix:///run/containerd/s/fd4140c1f6b01f5c4ef870d98460d6cdb279ef5ffe71f405e7cd39bcb14c44de" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:56.409421 systemd[1]: Started cri-containerd-390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3.scope - libcontainer container 390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3. Dec 12 18:39:56.461497 systemd-networkd[1447]: cali970d9e124fd: Link UP Dec 12 18:39:56.462771 systemd-networkd[1447]: cali970d9e124fd: Gained carrier Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.199 [INFO][4307] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.222 [INFO][4307] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0 calico-apiserver-59789c7c4b- calico-apiserver 986a62cf-f52b-49ec-a518-51f0449c9112 806 0 2025-12-12 18:39:29 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:59789c7c4b projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459-2-2-7-adcd881523 calico-apiserver-59789c7c4b-wht4c eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali970d9e124fd [] [] }} ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-wht4c" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.222 [INFO][4307] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-wht4c" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.295 [INFO][4347] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" HandleID="k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.295 [INFO][4347] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" HandleID="k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f890), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4459-2-2-7-adcd881523", "pod":"calico-apiserver-59789c7c4b-wht4c", "timestamp":"2025-12-12 18:39:56.295613237 +0000 UTC"}, Hostname:"ci-4459-2-2-7-adcd881523", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.296 [INFO][4347] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.338 [INFO][4347] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.338 [INFO][4347] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-2-2-7-adcd881523' Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.404 [INFO][4347] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.426 [INFO][4347] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.436 [INFO][4347] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.438 [INFO][4347] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.440 [INFO][4347] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.440 [INFO][4347] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.441 [INFO][4347] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711 Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.447 [INFO][4347] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.452 [INFO][4347] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.197/26] block=192.168.109.192/26 handle="k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.452 [INFO][4347] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.197/26] handle="k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.452 [INFO][4347] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 12 18:39:56.480142 containerd[1549]: 2025-12-12 18:39:56.452 [INFO][4347] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.197/26] IPv6=[] ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" HandleID="k8s-pod-network.6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" Dec 12 18:39:56.481405 containerd[1549]: 2025-12-12 18:39:56.458 [INFO][4307] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-wht4c" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0", GenerateName:"calico-apiserver-59789c7c4b-", Namespace:"calico-apiserver", SelfLink:"", UID:"986a62cf-f52b-49ec-a518-51f0449c9112", ResourceVersion:"806", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59789c7c4b", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"", Pod:"calico-apiserver-59789c7c4b-wht4c", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.109.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali970d9e124fd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:56.481405 containerd[1549]: 2025-12-12 18:39:56.458 [INFO][4307] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.197/32] ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-wht4c" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" Dec 12 18:39:56.481405 containerd[1549]: 2025-12-12 18:39:56.458 [INFO][4307] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali970d9e124fd ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-wht4c" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" Dec 12 18:39:56.481405 containerd[1549]: 2025-12-12 18:39:56.464 [INFO][4307] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-wht4c" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" Dec 12 18:39:56.481405 containerd[1549]: 2025-12-12 18:39:56.464 [INFO][4307] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-wht4c" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0", GenerateName:"calico-apiserver-59789c7c4b-", Namespace:"calico-apiserver", SelfLink:"", UID:"986a62cf-f52b-49ec-a518-51f0449c9112", ResourceVersion:"806", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59789c7c4b", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711", Pod:"calico-apiserver-59789c7c4b-wht4c", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.109.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali970d9e124fd", MAC:"fa:19:98:c5:59:20", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:56.481405 containerd[1549]: 2025-12-12 18:39:56.478 [INFO][4307] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-wht4c" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--wht4c-eth0" Dec 12 18:39:56.488672 containerd[1549]: time="2025-12-12T18:39:56.488632795Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7bbfbd7b8d-c5gdp,Uid:53428c7e-77fb-479f-978b-3877d65c8e58,Namespace:calico-system,Attempt:0,} returns sandbox id \"390b1c01ef35142221d929475b5b248d6b5e11ee138280eaccf1e01eec2971b3\"" Dec 12 18:39:56.490800 containerd[1549]: time="2025-12-12T18:39:56.490686753Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Dec 12 18:39:56.507870 containerd[1549]: time="2025-12-12T18:39:56.507836736Z" level=info msg="connecting to shim 6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711" address="unix:///run/containerd/s/b36c409e597fb7cb87f022f9a06825da461fff31cc6f2a7c723cd3a1295df61f" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:56.532388 systemd[1]: Started cri-containerd-6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711.scope - libcontainer container 6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711. Dec 12 18:39:56.566977 systemd-networkd[1447]: cali7f146dc0baf: Link UP Dec 12 18:39:56.567447 systemd-networkd[1447]: cali7f146dc0baf: Gained carrier Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.200 [INFO][4325] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.222 [INFO][4325] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0 calico-apiserver-59789c7c4b- calico-apiserver ae09f5bd-f7e2-460b-b94b-eb370fa59677 808 0 2025-12-12 18:39:29 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:59789c7c4b projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459-2-2-7-adcd881523 calico-apiserver-59789c7c4b-4fs8q eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali7f146dc0baf [] [] }} ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-4fs8q" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.223 [INFO][4325] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-4fs8q" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.295 [INFO][4351] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" HandleID="k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.296 [INFO][4351] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" HandleID="k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f8a0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4459-2-2-7-adcd881523", "pod":"calico-apiserver-59789c7c4b-4fs8q", "timestamp":"2025-12-12 18:39:56.295773567 +0000 UTC"}, Hostname:"ci-4459-2-2-7-adcd881523", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.296 [INFO][4351] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.454 [INFO][4351] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.454 [INFO][4351] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-2-2-7-adcd881523' Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.504 [INFO][4351] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.518 [INFO][4351] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.539 [INFO][4351] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.543 [INFO][4351] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.547 [INFO][4351] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.547 [INFO][4351] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.548 [INFO][4351] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05 Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.553 [INFO][4351] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.563 [INFO][4351] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.198/26] block=192.168.109.192/26 handle="k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.563 [INFO][4351] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.198/26] handle="k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.563 [INFO][4351] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 12 18:39:56.584630 containerd[1549]: 2025-12-12 18:39:56.563 [INFO][4351] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.198/26] IPv6=[] ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" HandleID="k8s-pod-network.79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Workload="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" Dec 12 18:39:56.585112 containerd[1549]: 2025-12-12 18:39:56.565 [INFO][4325] cni-plugin/k8s.go 418: Populated endpoint ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-4fs8q" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0", GenerateName:"calico-apiserver-59789c7c4b-", Namespace:"calico-apiserver", SelfLink:"", UID:"ae09f5bd-f7e2-460b-b94b-eb370fa59677", ResourceVersion:"808", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59789c7c4b", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"", Pod:"calico-apiserver-59789c7c4b-4fs8q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.109.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali7f146dc0baf", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:56.585112 containerd[1549]: 2025-12-12 18:39:56.565 [INFO][4325] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.198/32] ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-4fs8q" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" Dec 12 18:39:56.585112 containerd[1549]: 2025-12-12 18:39:56.565 [INFO][4325] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7f146dc0baf ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-4fs8q" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" Dec 12 18:39:56.585112 containerd[1549]: 2025-12-12 18:39:56.567 [INFO][4325] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-4fs8q" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" Dec 12 18:39:56.585112 containerd[1549]: 2025-12-12 18:39:56.567 [INFO][4325] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-4fs8q" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0", GenerateName:"calico-apiserver-59789c7c4b-", Namespace:"calico-apiserver", SelfLink:"", UID:"ae09f5bd-f7e2-460b-b94b-eb370fa59677", ResourceVersion:"808", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59789c7c4b", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05", Pod:"calico-apiserver-59789c7c4b-4fs8q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.109.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali7f146dc0baf", MAC:"7a:70:11:82:93:0f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:56.585112 containerd[1549]: 2025-12-12 18:39:56.582 [INFO][4325] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" Namespace="calico-apiserver" Pod="calico-apiserver-59789c7c4b-4fs8q" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-calico--apiserver--59789c7c4b--4fs8q-eth0" Dec 12 18:39:56.593433 containerd[1549]: time="2025-12-12T18:39:56.593395522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59789c7c4b-wht4c,Uid:986a62cf-f52b-49ec-a518-51f0449c9112,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"6008b3d5ada108c52581cf17f09e539d20236ba2518945905aeff36e7ef69711\"" Dec 12 18:39:56.607987 containerd[1549]: time="2025-12-12T18:39:56.607945252Z" level=info msg="connecting to shim 79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05" address="unix:///run/containerd/s/0775bf11a51962fa2e6c5cd223b90d10e7133d3f6a378e45a365beb4b2b1b21f" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:56.629314 systemd[1]: Started cri-containerd-79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05.scope - libcontainer container 79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05. Dec 12 18:39:56.668454 containerd[1549]: time="2025-12-12T18:39:56.668357785Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59789c7c4b-4fs8q,Uid:ae09f5bd-f7e2-460b-b94b-eb370fa59677,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"79c49105cc6694aa9306eb0af15bbfbe0e3709510516b8cde9b855d757117f05\"" Dec 12 18:39:56.910627 containerd[1549]: time="2025-12-12T18:39:56.910566124Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:39:56.911432 containerd[1549]: time="2025-12-12T18:39:56.911396048Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Dec 12 18:39:56.911491 containerd[1549]: time="2025-12-12T18:39:56.911461842Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Dec 12 18:39:56.911674 kubelet[2737]: E1212 18:39:56.911635 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 12 18:39:56.911767 kubelet[2737]: E1212 18:39:56.911680 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 12 18:39:56.911767 kubelet[2737]: E1212 18:39:56.911899 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mwjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-7bbfbd7b8d-c5gdp_calico-system(53428c7e-77fb-479f-978b-3877d65c8e58): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Dec 12 18:39:56.912821 containerd[1549]: time="2025-12-12T18:39:56.911934127Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 12 18:39:56.913594 kubelet[2737]: E1212 18:39:56.913554 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:39:57.119406 containerd[1549]: time="2025-12-12T18:39:57.119324681Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-f26wm,Uid:3431bd32-21c9-4310-bd4d-682f372a9f24,Namespace:kube-system,Attempt:0,}" Dec 12 18:39:57.229684 systemd-networkd[1447]: cali2cf33ddfed9: Link UP Dec 12 18:39:57.232298 systemd-networkd[1447]: cali2cf33ddfed9: Gained carrier Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.147 [INFO][4543] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.157 [INFO][4543] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0 coredns-674b8bbfcf- kube-system 3431bd32-21c9-4310-bd4d-682f372a9f24 805 0 2025-12-12 18:39:19 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4459-2-2-7-adcd881523 coredns-674b8bbfcf-f26wm eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali2cf33ddfed9 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Namespace="kube-system" Pod="coredns-674b8bbfcf-f26wm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.157 [INFO][4543] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Namespace="kube-system" Pod="coredns-674b8bbfcf-f26wm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.185 [INFO][4554] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" HandleID="k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Workload="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.185 [INFO][4554] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" HandleID="k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Workload="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f010), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4459-2-2-7-adcd881523", "pod":"coredns-674b8bbfcf-f26wm", "timestamp":"2025-12-12 18:39:57.185047098 +0000 UTC"}, Hostname:"ci-4459-2-2-7-adcd881523", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.185 [INFO][4554] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.185 [INFO][4554] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.185 [INFO][4554] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-2-2-7-adcd881523' Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.192 [INFO][4554] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.197 [INFO][4554] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.202 [INFO][4554] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.204 [INFO][4554] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.206 [INFO][4554] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.206 [INFO][4554] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.208 [INFO][4554] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2 Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.215 [INFO][4554] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.221 [INFO][4554] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.199/26] block=192.168.109.192/26 handle="k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.221 [INFO][4554] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.199/26] handle="k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.221 [INFO][4554] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 12 18:39:57.249268 containerd[1549]: 2025-12-12 18:39:57.221 [INFO][4554] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.199/26] IPv6=[] ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" HandleID="k8s-pod-network.25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Workload="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" Dec 12 18:39:57.250019 containerd[1549]: 2025-12-12 18:39:57.224 [INFO][4543] cni-plugin/k8s.go 418: Populated endpoint ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Namespace="kube-system" Pod="coredns-674b8bbfcf-f26wm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"3431bd32-21c9-4310-bd4d-682f372a9f24", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"", Pod:"coredns-674b8bbfcf-f26wm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.109.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2cf33ddfed9", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:57.250019 containerd[1549]: 2025-12-12 18:39:57.225 [INFO][4543] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.199/32] ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Namespace="kube-system" Pod="coredns-674b8bbfcf-f26wm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" Dec 12 18:39:57.250019 containerd[1549]: 2025-12-12 18:39:57.226 [INFO][4543] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2cf33ddfed9 ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Namespace="kube-system" Pod="coredns-674b8bbfcf-f26wm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" Dec 12 18:39:57.250019 containerd[1549]: 2025-12-12 18:39:57.231 [INFO][4543] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Namespace="kube-system" Pod="coredns-674b8bbfcf-f26wm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" Dec 12 18:39:57.250019 containerd[1549]: 2025-12-12 18:39:57.231 [INFO][4543] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Namespace="kube-system" Pod="coredns-674b8bbfcf-f26wm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"3431bd32-21c9-4310-bd4d-682f372a9f24", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2", Pod:"coredns-674b8bbfcf-f26wm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.109.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2cf33ddfed9", MAC:"c2:fb:c1:f6:56:a5", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:57.250019 containerd[1549]: 2025-12-12 18:39:57.244 [INFO][4543] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" Namespace="kube-system" Pod="coredns-674b8bbfcf-f26wm" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-coredns--674b8bbfcf--f26wm-eth0" Dec 12 18:39:57.271814 containerd[1549]: time="2025-12-12T18:39:57.271637194Z" level=info msg="connecting to shim 25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2" address="unix:///run/containerd/s/e7f93d132cafb1ab787546c2811f833967672c2060edd478c19802550c31fb99" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:57.296316 systemd[1]: Started cri-containerd-25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2.scope - libcontainer container 25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2. Dec 12 18:39:57.336032 containerd[1549]: time="2025-12-12T18:39:57.335978903Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:39:57.339217 containerd[1549]: time="2025-12-12T18:39:57.338719428Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 12 18:39:57.339217 containerd[1549]: time="2025-12-12T18:39:57.339166917Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Dec 12 18:39:57.339660 kubelet[2737]: E1212 18:39:57.339468 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:39:57.340890 kubelet[2737]: E1212 18:39:57.340012 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:39:57.340890 kubelet[2737]: E1212 18:39:57.340282 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lpjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59789c7c4b-wht4c_calico-apiserver(986a62cf-f52b-49ec-a518-51f0449c9112): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 12 18:39:57.341337 containerd[1549]: time="2025-12-12T18:39:57.340552072Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 12 18:39:57.341834 kubelet[2737]: E1212 18:39:57.341729 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:39:57.347430 containerd[1549]: time="2025-12-12T18:39:57.347405984Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-f26wm,Uid:3431bd32-21c9-4310-bd4d-682f372a9f24,Namespace:kube-system,Attempt:0,} returns sandbox id \"25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2\"" Dec 12 18:39:57.347632 kubelet[2737]: E1212 18:39:57.347587 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:39:57.353761 containerd[1549]: time="2025-12-12T18:39:57.353681873Z" level=info msg="CreateContainer within sandbox \"25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 12 18:39:57.364217 containerd[1549]: time="2025-12-12T18:39:57.364045872Z" level=info msg="Container 816930eadfd1abf85c467f1fc781468243022bad94d87432ea00e1bd5d748778: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:39:57.376081 containerd[1549]: time="2025-12-12T18:39:57.376002857Z" level=info msg="CreateContainer within sandbox \"25e877fe9eeb942c30c0f5bc1b4081af34057fef2531f92673049953e70989f2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"816930eadfd1abf85c467f1fc781468243022bad94d87432ea00e1bd5d748778\"" Dec 12 18:39:57.377927 containerd[1549]: time="2025-12-12T18:39:57.377881807Z" level=info msg="StartContainer for \"816930eadfd1abf85c467f1fc781468243022bad94d87432ea00e1bd5d748778\"" Dec 12 18:39:57.379215 containerd[1549]: time="2025-12-12T18:39:57.379143150Z" level=info msg="connecting to shim 816930eadfd1abf85c467f1fc781468243022bad94d87432ea00e1bd5d748778" address="unix:///run/containerd/s/e7f93d132cafb1ab787546c2811f833967672c2060edd478c19802550c31fb99" protocol=ttrpc version=3 Dec 12 18:39:57.398381 systemd[1]: Started cri-containerd-816930eadfd1abf85c467f1fc781468243022bad94d87432ea00e1bd5d748778.scope - libcontainer container 816930eadfd1abf85c467f1fc781468243022bad94d87432ea00e1bd5d748778. Dec 12 18:39:57.442177 containerd[1549]: time="2025-12-12T18:39:57.442142083Z" level=info msg="StartContainer for \"816930eadfd1abf85c467f1fc781468243022bad94d87432ea00e1bd5d748778\" returns successfully" Dec 12 18:39:57.714350 systemd-networkd[1447]: cali970d9e124fd: Gained IPv6LL Dec 12 18:39:57.783322 containerd[1549]: time="2025-12-12T18:39:57.783274833Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:39:57.784463 containerd[1549]: time="2025-12-12T18:39:57.784433214Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 12 18:39:57.784536 containerd[1549]: time="2025-12-12T18:39:57.784505409Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Dec 12 18:39:57.784677 kubelet[2737]: E1212 18:39:57.784636 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:39:57.784739 kubelet[2737]: E1212 18:39:57.784682 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:39:57.784877 kubelet[2737]: E1212 18:39:57.784832 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hfpwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59789c7c4b-4fs8q_calico-apiserver(ae09f5bd-f7e2-460b-b94b-eb370fa59677): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 12 18:39:57.786222 kubelet[2737]: E1212 18:39:57.786166 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:39:57.970349 systemd-networkd[1447]: cali7f146dc0baf: Gained IPv6LL Dec 12 18:39:58.118976 containerd[1549]: time="2025-12-12T18:39:58.118918518Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2msts,Uid:583f364c-28f5-44f4-853f-72900703e8a2,Namespace:calico-system,Attempt:0,}" Dec 12 18:39:58.233948 systemd-networkd[1447]: cali7bf0f146455: Link UP Dec 12 18:39:58.235395 systemd-networkd[1447]: cali7bf0f146455: Gained carrier Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.146 [INFO][4657] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.158 [INFO][4657] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0 csi-node-driver- calico-system 583f364c-28f5-44f4-853f-72900703e8a2 703 0 2025-12-12 18:39:33 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4459-2-2-7-adcd881523 csi-node-driver-2msts eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali7bf0f146455 [] [] }} ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Namespace="calico-system" Pod="csi-node-driver-2msts" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.158 [INFO][4657] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Namespace="calico-system" Pod="csi-node-driver-2msts" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.189 [INFO][4669] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" HandleID="k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Workload="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.189 [INFO][4669] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" HandleID="k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Workload="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f200), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459-2-2-7-adcd881523", "pod":"csi-node-driver-2msts", "timestamp":"2025-12-12 18:39:58.189369339 +0000 UTC"}, Hostname:"ci-4459-2-2-7-adcd881523", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.189 [INFO][4669] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.189 [INFO][4669] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.189 [INFO][4669] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-2-2-7-adcd881523' Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.197 [INFO][4669] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.202 [INFO][4669] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.206 [INFO][4669] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.209 [INFO][4669] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.212 [INFO][4669] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.212 [INFO][4669] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.213 [INFO][4669] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.218 [INFO][4669] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.227 [INFO][4669] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.200/26] block=192.168.109.192/26 handle="k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.227 [INFO][4669] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.200/26] handle="k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" host="ci-4459-2-2-7-adcd881523" Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.227 [INFO][4669] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 12 18:39:58.249685 containerd[1549]: 2025-12-12 18:39:58.227 [INFO][4669] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.200/26] IPv6=[] ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" HandleID="k8s-pod-network.1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Workload="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" Dec 12 18:39:58.251740 containerd[1549]: 2025-12-12 18:39:58.230 [INFO][4657] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Namespace="calico-system" Pod="csi-node-driver-2msts" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"583f364c-28f5-44f4-853f-72900703e8a2", ResourceVersion:"703", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"", Pod:"csi-node-driver-2msts", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.109.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali7bf0f146455", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:58.251740 containerd[1549]: 2025-12-12 18:39:58.230 [INFO][4657] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.200/32] ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Namespace="calico-system" Pod="csi-node-driver-2msts" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" Dec 12 18:39:58.251740 containerd[1549]: 2025-12-12 18:39:58.230 [INFO][4657] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7bf0f146455 ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Namespace="calico-system" Pod="csi-node-driver-2msts" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" Dec 12 18:39:58.251740 containerd[1549]: 2025-12-12 18:39:58.232 [INFO][4657] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Namespace="calico-system" Pod="csi-node-driver-2msts" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" Dec 12 18:39:58.251740 containerd[1549]: 2025-12-12 18:39:58.232 [INFO][4657] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Namespace="calico-system" Pod="csi-node-driver-2msts" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"583f364c-28f5-44f4-853f-72900703e8a2", ResourceVersion:"703", Generation:0, CreationTimestamp:time.Date(2025, time.December, 12, 18, 39, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-2-2-7-adcd881523", ContainerID:"1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd", Pod:"csi-node-driver-2msts", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.109.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali7bf0f146455", MAC:"06:ce:ea:c3:39:73", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 12 18:39:58.251740 containerd[1549]: 2025-12-12 18:39:58.243 [INFO][4657] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" Namespace="calico-system" Pod="csi-node-driver-2msts" WorkloadEndpoint="ci--4459--2--2--7--adcd881523-k8s-csi--node--driver--2msts-eth0" Dec 12 18:39:58.268561 containerd[1549]: time="2025-12-12T18:39:58.268532196Z" level=info msg="connecting to shim 1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd" address="unix:///run/containerd/s/85ebca02409a434983c1e0f1b9042ee156f86a36a7b06facef4481f380eb3db2" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:39:58.297399 systemd[1]: Started cri-containerd-1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd.scope - libcontainer container 1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd. Dec 12 18:39:58.320662 containerd[1549]: time="2025-12-12T18:39:58.320578110Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2msts,Uid:583f364c-28f5-44f4-853f-72900703e8a2,Namespace:calico-system,Attempt:0,} returns sandbox id \"1a66ec440959675d010390dd5f88227840376d176c43c2f2fb44cb933d52c5dd\"" Dec 12 18:39:58.323227 containerd[1549]: time="2025-12-12T18:39:58.323180306Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Dec 12 18:39:58.354750 systemd-networkd[1447]: cali682565a5b18: Gained IPv6LL Dec 12 18:39:58.374058 kubelet[2737]: E1212 18:39:58.374014 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:39:58.374618 kubelet[2737]: E1212 18:39:58.374583 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:39:58.374806 kubelet[2737]: E1212 18:39:58.374749 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:39:58.381549 kubelet[2737]: I1212 18:39:58.381511 2737 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-f26wm" podStartSLOduration=39.381500918 podStartE2EDuration="39.381500918s" podCreationTimestamp="2025-12-12 18:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:39:58.380234204 +0000 UTC m=+45.359080835" watchObservedRunningTime="2025-12-12 18:39:58.381500918 +0000 UTC m=+45.360347517" Dec 12 18:39:58.762305 containerd[1549]: time="2025-12-12T18:39:58.762234461Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:39:58.764095 containerd[1549]: time="2025-12-12T18:39:58.763920370Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Dec 12 18:39:58.764095 containerd[1549]: time="2025-12-12T18:39:58.764000821Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Dec 12 18:39:58.764619 kubelet[2737]: E1212 18:39:58.764542 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 12 18:39:58.764619 kubelet[2737]: E1212 18:39:58.764608 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 12 18:39:58.764978 kubelet[2737]: E1212 18:39:58.764827 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cqz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Dec 12 18:39:58.768344 containerd[1549]: time="2025-12-12T18:39:58.768290188Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Dec 12 18:39:59.250519 systemd-networkd[1447]: cali2cf33ddfed9: Gained IPv6LL Dec 12 18:39:59.396118 containerd[1549]: time="2025-12-12T18:39:59.396017413Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:39:59.397953 containerd[1549]: time="2025-12-12T18:39:59.397849025Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Dec 12 18:39:59.398582 containerd[1549]: time="2025-12-12T18:39:59.397966986Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Dec 12 18:39:59.398680 kubelet[2737]: E1212 18:39:59.398230 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 12 18:39:59.398680 kubelet[2737]: E1212 18:39:59.398295 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 12 18:39:59.398680 kubelet[2737]: E1212 18:39:59.398488 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cqz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Dec 12 18:39:59.399873 kubelet[2737]: E1212 18:39:59.399674 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:39:59.763692 systemd-networkd[1447]: cali7bf0f146455: Gained IPv6LL Dec 12 18:40:00.363998 kubelet[2737]: E1212 18:40:00.363860 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:40:01.444935 kubelet[2737]: I1212 18:40:01.444651 2737 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 18:40:02.291406 kubelet[2737]: I1212 18:40:02.291130 2737 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 18:40:03.259425 systemd-networkd[1447]: vxlan.calico: Link UP Dec 12 18:40:03.259432 systemd-networkd[1447]: vxlan.calico: Gained carrier Dec 12 18:40:04.120394 containerd[1549]: time="2025-12-12T18:40:04.120349876Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Dec 12 18:40:04.541556 containerd[1549]: time="2025-12-12T18:40:04.541512098Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:04.542694 containerd[1549]: time="2025-12-12T18:40:04.542655792Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Dec 12 18:40:04.542844 containerd[1549]: time="2025-12-12T18:40:04.542732345Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Dec 12 18:40:04.542930 kubelet[2737]: E1212 18:40:04.542884 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 12 18:40:04.543460 kubelet[2737]: E1212 18:40:04.542935 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 12 18:40:04.543460 kubelet[2737]: E1212 18:40:04.543051 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:085cdf68872848a99fb533c1a7535d78,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-86lgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-774c4fc74b-q78zm_calico-system(48676334-6f17-4d47-84e1-7fe85ea764df): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:04.545183 containerd[1549]: time="2025-12-12T18:40:04.545123306Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Dec 12 18:40:04.997275 containerd[1549]: time="2025-12-12T18:40:04.997211994Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:04.998364 containerd[1549]: time="2025-12-12T18:40:04.998317445Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Dec 12 18:40:04.998521 containerd[1549]: time="2025-12-12T18:40:04.998398406Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Dec 12 18:40:04.998632 kubelet[2737]: E1212 18:40:04.998580 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 12 18:40:04.998685 kubelet[2737]: E1212 18:40:04.998633 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 12 18:40:04.999182 kubelet[2737]: E1212 18:40:04.998783 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86lgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-774c4fc74b-q78zm_calico-system(48676334-6f17-4d47-84e1-7fe85ea764df): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:05.000581 kubelet[2737]: E1212 18:40:05.000492 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:40:05.074464 systemd-networkd[1447]: vxlan.calico: Gained IPv6LL Dec 12 18:40:09.120662 containerd[1549]: time="2025-12-12T18:40:09.120620278Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Dec 12 18:40:09.535493 containerd[1549]: time="2025-12-12T18:40:09.535432300Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:09.536841 containerd[1549]: time="2025-12-12T18:40:09.536768304Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Dec 12 18:40:09.536937 containerd[1549]: time="2025-12-12T18:40:09.536872409Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Dec 12 18:40:09.537180 kubelet[2737]: E1212 18:40:09.537131 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 12 18:40:09.537473 kubelet[2737]: E1212 18:40:09.537182 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 12 18:40:09.538269 containerd[1549]: time="2025-12-12T18:40:09.537945510Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Dec 12 18:40:09.538484 kubelet[2737]: E1212 18:40:09.537882 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mwjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-7bbfbd7b8d-c5gdp_calico-system(53428c7e-77fb-479f-978b-3877d65c8e58): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:09.539621 kubelet[2737]: E1212 18:40:09.539591 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:40:10.005242 containerd[1549]: time="2025-12-12T18:40:10.004535609Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:10.006888 containerd[1549]: time="2025-12-12T18:40:10.006853844Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Dec 12 18:40:10.007861 kubelet[2737]: E1212 18:40:10.007780 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 12 18:40:10.007938 containerd[1549]: time="2025-12-12T18:40:10.007599863Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Dec 12 18:40:10.008073 kubelet[2737]: E1212 18:40:10.008024 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 12 18:40:10.008380 kubelet[2737]: E1212 18:40:10.008324 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5d5ds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-92gml_calico-system(3a064600-3d57-4e0d-80c2-a8041300095d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:10.010408 kubelet[2737]: E1212 18:40:10.010334 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:40:10.119963 containerd[1549]: time="2025-12-12T18:40:10.119852521Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 12 18:40:10.562941 containerd[1549]: time="2025-12-12T18:40:10.562879294Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:10.564239 containerd[1549]: time="2025-12-12T18:40:10.564199162Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 12 18:40:10.564372 containerd[1549]: time="2025-12-12T18:40:10.564222956Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Dec 12 18:40:10.564540 kubelet[2737]: E1212 18:40:10.564499 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:40:10.564755 kubelet[2737]: E1212 18:40:10.564560 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:40:10.564865 kubelet[2737]: E1212 18:40:10.564767 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hfpwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59789c7c4b-4fs8q_calico-apiserver(ae09f5bd-f7e2-460b-b94b-eb370fa59677): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:10.566498 kubelet[2737]: E1212 18:40:10.566460 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:40:11.122387 containerd[1549]: time="2025-12-12T18:40:11.122343238Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Dec 12 18:40:11.573257 containerd[1549]: time="2025-12-12T18:40:11.573205076Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:11.574618 containerd[1549]: time="2025-12-12T18:40:11.574563692Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Dec 12 18:40:11.574793 containerd[1549]: time="2025-12-12T18:40:11.574655980Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Dec 12 18:40:11.574858 kubelet[2737]: E1212 18:40:11.574804 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 12 18:40:11.574858 kubelet[2737]: E1212 18:40:11.574848 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 12 18:40:11.575144 kubelet[2737]: E1212 18:40:11.574995 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cqz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:11.577091 containerd[1549]: time="2025-12-12T18:40:11.577060369Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Dec 12 18:40:12.012598 containerd[1549]: time="2025-12-12T18:40:12.012425988Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:12.013831 containerd[1549]: time="2025-12-12T18:40:12.013772995Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Dec 12 18:40:12.013831 containerd[1549]: time="2025-12-12T18:40:12.013805404Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Dec 12 18:40:12.014023 kubelet[2737]: E1212 18:40:12.013961 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 12 18:40:12.014097 kubelet[2737]: E1212 18:40:12.014023 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 12 18:40:12.014183 kubelet[2737]: E1212 18:40:12.014138 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cqz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:12.015395 kubelet[2737]: E1212 18:40:12.015365 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:40:12.121114 containerd[1549]: time="2025-12-12T18:40:12.120721959Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 12 18:40:12.550077 containerd[1549]: time="2025-12-12T18:40:12.550022911Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:12.551501 containerd[1549]: time="2025-12-12T18:40:12.551438763Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 12 18:40:12.551591 containerd[1549]: time="2025-12-12T18:40:12.551533207Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Dec 12 18:40:12.551724 kubelet[2737]: E1212 18:40:12.551684 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:40:12.551832 kubelet[2737]: E1212 18:40:12.551759 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:40:12.551978 kubelet[2737]: E1212 18:40:12.551893 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lpjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59789c7c4b-wht4c_calico-apiserver(986a62cf-f52b-49ec-a518-51f0449c9112): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:12.553554 kubelet[2737]: E1212 18:40:12.553515 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:40:18.120734 kubelet[2737]: E1212 18:40:18.120660 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:40:22.119395 kubelet[2737]: E1212 18:40:22.119257 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:40:22.121422 kubelet[2737]: E1212 18:40:22.121327 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:40:23.120383 kubelet[2737]: E1212 18:40:23.119644 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:40:23.122107 kubelet[2737]: E1212 18:40:23.122064 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:40:26.120699 kubelet[2737]: E1212 18:40:26.120620 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:40:32.119617 containerd[1549]: time="2025-12-12T18:40:32.119051702Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Dec 12 18:40:32.564123 containerd[1549]: time="2025-12-12T18:40:32.564036772Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:32.565596 containerd[1549]: time="2025-12-12T18:40:32.565538249Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Dec 12 18:40:32.565775 containerd[1549]: time="2025-12-12T18:40:32.565631351Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Dec 12 18:40:32.565890 kubelet[2737]: E1212 18:40:32.565830 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 12 18:40:32.566352 kubelet[2737]: E1212 18:40:32.565885 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 12 18:40:32.566352 kubelet[2737]: E1212 18:40:32.566032 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:085cdf68872848a99fb533c1a7535d78,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-86lgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-774c4fc74b-q78zm_calico-system(48676334-6f17-4d47-84e1-7fe85ea764df): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:32.568861 containerd[1549]: time="2025-12-12T18:40:32.568797295Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Dec 12 18:40:32.993959 containerd[1549]: time="2025-12-12T18:40:32.993908154Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:32.995503 containerd[1549]: time="2025-12-12T18:40:32.995339181Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Dec 12 18:40:32.995574 containerd[1549]: time="2025-12-12T18:40:32.995537207Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Dec 12 18:40:32.995741 kubelet[2737]: E1212 18:40:32.995673 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 12 18:40:32.995741 kubelet[2737]: E1212 18:40:32.995729 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 12 18:40:32.996407 kubelet[2737]: E1212 18:40:32.996290 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86lgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-774c4fc74b-q78zm_calico-system(48676334-6f17-4d47-84e1-7fe85ea764df): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:32.997491 kubelet[2737]: E1212 18:40:32.997447 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:40:33.122164 containerd[1549]: time="2025-12-12T18:40:33.121746158Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Dec 12 18:40:33.555217 containerd[1549]: time="2025-12-12T18:40:33.555007060Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:33.556221 containerd[1549]: time="2025-12-12T18:40:33.556172876Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Dec 12 18:40:33.556322 containerd[1549]: time="2025-12-12T18:40:33.556209053Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Dec 12 18:40:33.556453 kubelet[2737]: E1212 18:40:33.556404 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 12 18:40:33.556515 kubelet[2737]: E1212 18:40:33.556451 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 12 18:40:33.556606 kubelet[2737]: E1212 18:40:33.556568 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5d5ds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-92gml_calico-system(3a064600-3d57-4e0d-80c2-a8041300095d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:33.558273 kubelet[2737]: E1212 18:40:33.558248 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:40:35.123269 containerd[1549]: time="2025-12-12T18:40:35.122637057Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Dec 12 18:40:35.565668 containerd[1549]: time="2025-12-12T18:40:35.565514858Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:35.567212 containerd[1549]: time="2025-12-12T18:40:35.567115822Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Dec 12 18:40:35.567614 containerd[1549]: time="2025-12-12T18:40:35.567173077Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Dec 12 18:40:35.567803 kubelet[2737]: E1212 18:40:35.567735 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 12 18:40:35.567803 kubelet[2737]: E1212 18:40:35.567779 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 12 18:40:35.568509 kubelet[2737]: E1212 18:40:35.567951 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cqz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:35.568800 containerd[1549]: time="2025-12-12T18:40:35.568739307Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Dec 12 18:40:36.000312 containerd[1549]: time="2025-12-12T18:40:36.000242851Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:36.001486 containerd[1549]: time="2025-12-12T18:40:36.001421534Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Dec 12 18:40:36.001658 containerd[1549]: time="2025-12-12T18:40:36.001505780Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Dec 12 18:40:36.002093 kubelet[2737]: E1212 18:40:36.001826 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 12 18:40:36.002093 kubelet[2737]: E1212 18:40:36.001874 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 12 18:40:36.002477 kubelet[2737]: E1212 18:40:36.002249 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mwjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-7bbfbd7b8d-c5gdp_calico-system(53428c7e-77fb-479f-978b-3877d65c8e58): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:36.003228 containerd[1549]: time="2025-12-12T18:40:36.002644920Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Dec 12 18:40:36.003780 kubelet[2737]: E1212 18:40:36.003348 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:40:36.442969 containerd[1549]: time="2025-12-12T18:40:36.442736302Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:36.444579 containerd[1549]: time="2025-12-12T18:40:36.444456567Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Dec 12 18:40:36.444811 containerd[1549]: time="2025-12-12T18:40:36.444782220Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Dec 12 18:40:36.444908 kubelet[2737]: E1212 18:40:36.444775 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 12 18:40:36.444908 kubelet[2737]: E1212 18:40:36.444821 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 12 18:40:36.445058 kubelet[2737]: E1212 18:40:36.444956 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cqz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:36.446549 kubelet[2737]: E1212 18:40:36.446497 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:40:37.122979 containerd[1549]: time="2025-12-12T18:40:37.121665547Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 12 18:40:37.568654 containerd[1549]: time="2025-12-12T18:40:37.568603915Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:37.570037 containerd[1549]: time="2025-12-12T18:40:37.569955118Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 12 18:40:37.570144 containerd[1549]: time="2025-12-12T18:40:37.570060993Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Dec 12 18:40:37.570348 kubelet[2737]: E1212 18:40:37.570303 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:40:37.570639 kubelet[2737]: E1212 18:40:37.570354 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:40:37.571123 kubelet[2737]: E1212 18:40:37.570649 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lpjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59789c7c4b-wht4c_calico-apiserver(986a62cf-f52b-49ec-a518-51f0449c9112): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:37.572239 kubelet[2737]: E1212 18:40:37.572182 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:40:38.122131 containerd[1549]: time="2025-12-12T18:40:38.122095266Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 12 18:40:38.574764 containerd[1549]: time="2025-12-12T18:40:38.574703938Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:40:38.576244 containerd[1549]: time="2025-12-12T18:40:38.576138716Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 12 18:40:38.576366 containerd[1549]: time="2025-12-12T18:40:38.576204329Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Dec 12 18:40:38.576496 kubelet[2737]: E1212 18:40:38.576440 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:40:38.576496 kubelet[2737]: E1212 18:40:38.576491 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:40:38.576810 kubelet[2737]: E1212 18:40:38.576629 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hfpwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59789c7c4b-4fs8q_calico-apiserver(ae09f5bd-f7e2-460b-b94b-eb370fa59677): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 12 18:40:38.578283 kubelet[2737]: E1212 18:40:38.578217 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:40:46.122164 kubelet[2737]: E1212 18:40:46.122081 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:40:47.122464 kubelet[2737]: E1212 18:40:47.122417 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:40:47.123316 kubelet[2737]: E1212 18:40:47.123069 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:40:49.122940 kubelet[2737]: E1212 18:40:49.122754 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:40:50.120866 kubelet[2737]: E1212 18:40:50.120807 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:40:50.121277 kubelet[2737]: E1212 18:40:50.121216 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:40:58.123944 kubelet[2737]: E1212 18:40:58.123720 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:40:59.121825 kubelet[2737]: E1212 18:40:59.121764 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:40:59.123699 kubelet[2737]: E1212 18:40:59.123365 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:41:01.122304 kubelet[2737]: E1212 18:41:01.122045 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:41:02.126025 kubelet[2737]: E1212 18:41:02.124949 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:41:02.127048 kubelet[2737]: E1212 18:41:02.126864 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:41:03.711438 systemd[1]: Started sshd@7-65.21.248.196:22-147.75.109.163:42530.service - OpenSSH per-connection server daemon (147.75.109.163:42530). Dec 12 18:41:04.870896 sshd[5099]: Accepted publickey for core from 147.75.109.163 port 42530 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:04.875724 sshd-session[5099]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:04.884866 systemd-logind[1520]: New session 8 of user core. Dec 12 18:41:04.889506 systemd[1]: Started session-8.scope - Session 8 of User core. Dec 12 18:41:06.218768 sshd[5103]: Connection closed by 147.75.109.163 port 42530 Dec 12 18:41:06.220364 sshd-session[5099]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:06.224690 systemd-logind[1520]: Session 8 logged out. Waiting for processes to exit. Dec 12 18:41:06.226470 systemd[1]: sshd@7-65.21.248.196:22-147.75.109.163:42530.service: Deactivated successfully. Dec 12 18:41:06.228970 systemd[1]: session-8.scope: Deactivated successfully. Dec 12 18:41:06.231629 systemd-logind[1520]: Removed session 8. Dec 12 18:41:10.121682 kubelet[2737]: E1212 18:41:10.121609 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:41:11.383673 systemd[1]: Started sshd@8-65.21.248.196:22-147.75.109.163:42534.service - OpenSSH per-connection server daemon (147.75.109.163:42534). Dec 12 18:41:12.126580 kubelet[2737]: E1212 18:41:12.126533 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:41:12.416928 sshd[5117]: Accepted publickey for core from 147.75.109.163 port 42534 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:12.420721 sshd-session[5117]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:12.429120 systemd-logind[1520]: New session 9 of user core. Dec 12 18:41:12.432710 systemd[1]: Started session-9.scope - Session 9 of User core. Dec 12 18:41:13.132468 kubelet[2737]: E1212 18:41:13.132400 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:41:13.309785 sshd[5120]: Connection closed by 147.75.109.163 port 42534 Dec 12 18:41:13.311450 sshd-session[5117]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:13.317873 systemd[1]: sshd@8-65.21.248.196:22-147.75.109.163:42534.service: Deactivated successfully. Dec 12 18:41:13.324832 systemd[1]: session-9.scope: Deactivated successfully. Dec 12 18:41:13.330970 systemd-logind[1520]: Session 9 logged out. Waiting for processes to exit. Dec 12 18:41:13.333140 systemd-logind[1520]: Removed session 9. Dec 12 18:41:13.488372 systemd[1]: Started sshd@9-65.21.248.196:22-147.75.109.163:39078.service - OpenSSH per-connection server daemon (147.75.109.163:39078). Dec 12 18:41:14.119306 containerd[1549]: time="2025-12-12T18:41:14.119101696Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Dec 12 18:41:14.121177 kubelet[2737]: E1212 18:41:14.121132 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:41:14.493964 sshd[5135]: Accepted publickey for core from 147.75.109.163 port 39078 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:14.496489 sshd-session[5135]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:14.502232 systemd-logind[1520]: New session 10 of user core. Dec 12 18:41:14.505558 systemd[1]: Started session-10.scope - Session 10 of User core. Dec 12 18:41:14.560205 containerd[1549]: time="2025-12-12T18:41:14.560114762Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:41:14.561988 containerd[1549]: time="2025-12-12T18:41:14.561959111Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Dec 12 18:41:14.562107 containerd[1549]: time="2025-12-12T18:41:14.562024563Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Dec 12 18:41:14.562215 kubelet[2737]: E1212 18:41:14.562157 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 12 18:41:14.563299 kubelet[2737]: E1212 18:41:14.562239 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 12 18:41:14.563299 kubelet[2737]: E1212 18:41:14.562688 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5d5ds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-92gml_calico-system(3a064600-3d57-4e0d-80c2-a8041300095d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Dec 12 18:41:14.563869 kubelet[2737]: E1212 18:41:14.563842 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:41:15.122567 kubelet[2737]: E1212 18:41:15.122494 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:41:15.315441 sshd[5138]: Connection closed by 147.75.109.163 port 39078 Dec 12 18:41:15.316388 sshd-session[5135]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:15.323959 systemd-logind[1520]: Session 10 logged out. Waiting for processes to exit. Dec 12 18:41:15.324585 systemd[1]: sshd@9-65.21.248.196:22-147.75.109.163:39078.service: Deactivated successfully. Dec 12 18:41:15.328706 systemd[1]: session-10.scope: Deactivated successfully. Dec 12 18:41:15.332944 systemd-logind[1520]: Removed session 10. Dec 12 18:41:15.521378 systemd[1]: Started sshd@10-65.21.248.196:22-147.75.109.163:39084.service - OpenSSH per-connection server daemon (147.75.109.163:39084). Dec 12 18:41:16.622991 sshd[5149]: Accepted publickey for core from 147.75.109.163 port 39084 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:16.625106 sshd-session[5149]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:16.631112 systemd-logind[1520]: New session 11 of user core. Dec 12 18:41:16.636320 systemd[1]: Started session-11.scope - Session 11 of User core. Dec 12 18:41:17.473676 sshd[5152]: Connection closed by 147.75.109.163 port 39084 Dec 12 18:41:17.476248 sshd-session[5149]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:17.481007 systemd-logind[1520]: Session 11 logged out. Waiting for processes to exit. Dec 12 18:41:17.483062 systemd[1]: sshd@10-65.21.248.196:22-147.75.109.163:39084.service: Deactivated successfully. Dec 12 18:41:17.487947 systemd[1]: session-11.scope: Deactivated successfully. Dec 12 18:41:17.490996 systemd-logind[1520]: Removed session 11. Dec 12 18:41:22.121714 containerd[1549]: time="2025-12-12T18:41:22.121328643Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Dec 12 18:41:22.553572 containerd[1549]: time="2025-12-12T18:41:22.553450823Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:41:22.555438 containerd[1549]: time="2025-12-12T18:41:22.555325621Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Dec 12 18:41:22.555561 containerd[1549]: time="2025-12-12T18:41:22.555522159Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Dec 12 18:41:22.556145 kubelet[2737]: E1212 18:41:22.556035 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 12 18:41:22.556655 kubelet[2737]: E1212 18:41:22.556171 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 12 18:41:22.557221 kubelet[2737]: E1212 18:41:22.557073 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cqz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Dec 12 18:41:22.559965 containerd[1549]: time="2025-12-12T18:41:22.559899415Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Dec 12 18:41:22.630696 systemd[1]: Started sshd@11-65.21.248.196:22-147.75.109.163:55360.service - OpenSSH per-connection server daemon (147.75.109.163:55360). Dec 12 18:41:22.999285 containerd[1549]: time="2025-12-12T18:41:22.999156598Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:41:23.000963 containerd[1549]: time="2025-12-12T18:41:23.000864064Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Dec 12 18:41:23.001838 containerd[1549]: time="2025-12-12T18:41:23.000975091Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Dec 12 18:41:23.001885 kubelet[2737]: E1212 18:41:23.001180 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 12 18:41:23.001885 kubelet[2737]: E1212 18:41:23.001263 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 12 18:41:23.001885 kubelet[2737]: E1212 18:41:23.001783 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cqz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2msts_calico-system(583f364c-28f5-44f4-853f-72900703e8a2): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Dec 12 18:41:23.002999 kubelet[2737]: E1212 18:41:23.002951 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:41:23.633039 sshd[5172]: Accepted publickey for core from 147.75.109.163 port 55360 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:23.635743 sshd-session[5172]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:23.645644 systemd-logind[1520]: New session 12 of user core. Dec 12 18:41:23.650575 systemd[1]: Started session-12.scope - Session 12 of User core. Dec 12 18:41:24.124246 containerd[1549]: time="2025-12-12T18:41:24.123592617Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Dec 12 18:41:24.466096 sshd[5175]: Connection closed by 147.75.109.163 port 55360 Dec 12 18:41:24.467048 sshd-session[5172]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:24.473902 systemd-logind[1520]: Session 12 logged out. Waiting for processes to exit. Dec 12 18:41:24.474278 systemd[1]: sshd@11-65.21.248.196:22-147.75.109.163:55360.service: Deactivated successfully. Dec 12 18:41:24.476725 systemd[1]: session-12.scope: Deactivated successfully. Dec 12 18:41:24.479956 systemd-logind[1520]: Removed session 12. Dec 12 18:41:24.559621 containerd[1549]: time="2025-12-12T18:41:24.559544775Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:41:24.561118 containerd[1549]: time="2025-12-12T18:41:24.560997377Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Dec 12 18:41:24.561873 containerd[1549]: time="2025-12-12T18:41:24.561151525Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Dec 12 18:41:24.561944 kubelet[2737]: E1212 18:41:24.561428 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 12 18:41:24.561944 kubelet[2737]: E1212 18:41:24.561526 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 12 18:41:24.561944 kubelet[2737]: E1212 18:41:24.561739 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:085cdf68872848a99fb533c1a7535d78,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-86lgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-774c4fc74b-q78zm_calico-system(48676334-6f17-4d47-84e1-7fe85ea764df): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Dec 12 18:41:24.566151 containerd[1549]: time="2025-12-12T18:41:24.565937486Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Dec 12 18:41:24.978110 containerd[1549]: time="2025-12-12T18:41:24.978051151Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:41:24.979123 containerd[1549]: time="2025-12-12T18:41:24.979079490Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Dec 12 18:41:24.979222 containerd[1549]: time="2025-12-12T18:41:24.979160160Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Dec 12 18:41:24.980418 kubelet[2737]: E1212 18:41:24.980363 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 12 18:41:24.980486 kubelet[2737]: E1212 18:41:24.980438 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 12 18:41:24.981201 kubelet[2737]: E1212 18:41:24.980856 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86lgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-774c4fc74b-q78zm_calico-system(48676334-6f17-4d47-84e1-7fe85ea764df): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Dec 12 18:41:24.982066 kubelet[2737]: E1212 18:41:24.982014 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:41:25.123090 containerd[1549]: time="2025-12-12T18:41:25.121885129Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 12 18:41:25.549571 containerd[1549]: time="2025-12-12T18:41:25.549510493Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:41:25.550713 containerd[1549]: time="2025-12-12T18:41:25.550668314Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 12 18:41:25.550767 containerd[1549]: time="2025-12-12T18:41:25.550754946Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Dec 12 18:41:25.551830 kubelet[2737]: E1212 18:41:25.551782 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:41:25.551885 kubelet[2737]: E1212 18:41:25.551857 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:41:25.552997 kubelet[2737]: E1212 18:41:25.552327 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lpjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59789c7c4b-wht4c_calico-apiserver(986a62cf-f52b-49ec-a518-51f0449c9112): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 12 18:41:25.553511 kubelet[2737]: E1212 18:41:25.553464 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:41:26.124023 containerd[1549]: time="2025-12-12T18:41:26.123706730Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 12 18:41:26.126334 kubelet[2737]: E1212 18:41:26.124997 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:41:26.566356 containerd[1549]: time="2025-12-12T18:41:26.566310090Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:41:26.567565 containerd[1549]: time="2025-12-12T18:41:26.567503988Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 12 18:41:26.567946 containerd[1549]: time="2025-12-12T18:41:26.567622310Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Dec 12 18:41:26.567976 kubelet[2737]: E1212 18:41:26.567830 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:41:26.567976 kubelet[2737]: E1212 18:41:26.567889 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 12 18:41:26.568129 kubelet[2737]: E1212 18:41:26.568079 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hfpwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59789c7c4b-4fs8q_calico-apiserver(ae09f5bd-f7e2-460b-b94b-eb370fa59677): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 12 18:41:26.568833 containerd[1549]: time="2025-12-12T18:41:26.568766084Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Dec 12 18:41:26.570155 kubelet[2737]: E1212 18:41:26.570123 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:41:27.009065 containerd[1549]: time="2025-12-12T18:41:27.008892542Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 12 18:41:27.010209 containerd[1549]: time="2025-12-12T18:41:27.010072024Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Dec 12 18:41:27.010209 containerd[1549]: time="2025-12-12T18:41:27.010143236Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Dec 12 18:41:27.010470 kubelet[2737]: E1212 18:41:27.010438 2737 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 12 18:41:27.010688 kubelet[2737]: E1212 18:41:27.010647 2737 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 12 18:41:27.010938 kubelet[2737]: E1212 18:41:27.010890 2737 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mwjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-7bbfbd7b8d-c5gdp_calico-system(53428c7e-77fb-479f-978b-3877d65c8e58): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Dec 12 18:41:27.012516 kubelet[2737]: E1212 18:41:27.012476 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:41:29.672431 systemd[1]: Started sshd@12-65.21.248.196:22-147.75.109.163:55368.service - OpenSSH per-connection server daemon (147.75.109.163:55368). Dec 12 18:41:30.798564 sshd[5196]: Accepted publickey for core from 147.75.109.163 port 55368 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:30.800886 sshd-session[5196]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:30.808604 systemd-logind[1520]: New session 13 of user core. Dec 12 18:41:30.815719 systemd[1]: Started session-13.scope - Session 13 of User core. Dec 12 18:41:31.641315 sshd[5199]: Connection closed by 147.75.109.163 port 55368 Dec 12 18:41:31.642865 sshd-session[5196]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:31.647326 systemd-logind[1520]: Session 13 logged out. Waiting for processes to exit. Dec 12 18:41:31.648309 systemd[1]: sshd@12-65.21.248.196:22-147.75.109.163:55368.service: Deactivated successfully. Dec 12 18:41:31.652121 systemd[1]: session-13.scope: Deactivated successfully. Dec 12 18:41:31.657334 systemd-logind[1520]: Removed session 13. Dec 12 18:41:31.826583 systemd[1]: Started sshd@13-65.21.248.196:22-147.75.109.163:55372.service - OpenSSH per-connection server daemon (147.75.109.163:55372). Dec 12 18:41:32.916820 sshd[5234]: Accepted publickey for core from 147.75.109.163 port 55372 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:32.917720 sshd-session[5234]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:32.923004 systemd-logind[1520]: New session 14 of user core. Dec 12 18:41:32.929098 systemd[1]: Started session-14.scope - Session 14 of User core. Dec 12 18:41:33.978979 sshd[5237]: Connection closed by 147.75.109.163 port 55372 Dec 12 18:41:33.986770 sshd-session[5234]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:33.995416 systemd-logind[1520]: Session 14 logged out. Waiting for processes to exit. Dec 12 18:41:33.995587 systemd[1]: sshd@13-65.21.248.196:22-147.75.109.163:55372.service: Deactivated successfully. Dec 12 18:41:33.997977 systemd[1]: session-14.scope: Deactivated successfully. Dec 12 18:41:33.999623 systemd-logind[1520]: Removed session 14. Dec 12 18:41:34.128230 systemd[1]: Started sshd@14-65.21.248.196:22-147.75.109.163:47516.service - OpenSSH per-connection server daemon (147.75.109.163:47516). Dec 12 18:41:35.123507 kubelet[2737]: E1212 18:41:35.123421 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:41:35.160660 sshd[5254]: Accepted publickey for core from 147.75.109.163 port 47516 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:35.162766 sshd-session[5254]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:35.169239 systemd-logind[1520]: New session 15 of user core. Dec 12 18:41:35.173353 systemd[1]: Started session-15.scope - Session 15 of User core. Dec 12 18:41:36.635641 sshd[5257]: Connection closed by 147.75.109.163 port 47516 Dec 12 18:41:36.636482 sshd-session[5254]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:36.642621 systemd[1]: sshd@14-65.21.248.196:22-147.75.109.163:47516.service: Deactivated successfully. Dec 12 18:41:36.645088 systemd[1]: session-15.scope: Deactivated successfully. Dec 12 18:41:36.646696 systemd-logind[1520]: Session 15 logged out. Waiting for processes to exit. Dec 12 18:41:36.648562 systemd-logind[1520]: Removed session 15. Dec 12 18:41:36.841566 systemd[1]: Started sshd@15-65.21.248.196:22-147.75.109.163:47528.service - OpenSSH per-connection server daemon (147.75.109.163:47528). Dec 12 18:41:37.953403 sshd[5288]: Accepted publickey for core from 147.75.109.163 port 47528 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:37.954915 sshd-session[5288]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:37.959660 systemd-logind[1520]: New session 16 of user core. Dec 12 18:41:37.965309 systemd[1]: Started session-16.scope - Session 16 of User core. Dec 12 18:41:38.119630 kubelet[2737]: E1212 18:41:38.119551 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:41:39.063399 sshd[5291]: Connection closed by 147.75.109.163 port 47528 Dec 12 18:41:39.064003 sshd-session[5288]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:39.077077 systemd[1]: sshd@15-65.21.248.196:22-147.75.109.163:47528.service: Deactivated successfully. Dec 12 18:41:39.083448 systemd[1]: session-16.scope: Deactivated successfully. Dec 12 18:41:39.086446 systemd-logind[1520]: Session 16 logged out. Waiting for processes to exit. Dec 12 18:41:39.090652 systemd-logind[1520]: Removed session 16. Dec 12 18:41:39.127557 kubelet[2737]: E1212 18:41:39.126071 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:41:39.128973 kubelet[2737]: E1212 18:41:39.124662 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d" Dec 12 18:41:39.129328 kubelet[2737]: E1212 18:41:39.128696 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:41:39.252362 systemd[1]: Started sshd@16-65.21.248.196:22-147.75.109.163:47530.service - OpenSSH per-connection server daemon (147.75.109.163:47530). Dec 12 18:41:40.125854 kubelet[2737]: E1212 18:41:40.125775 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-774c4fc74b-q78zm" podUID="48676334-6f17-4d47-84e1-7fe85ea764df" Dec 12 18:41:40.363916 sshd[5301]: Accepted publickey for core from 147.75.109.163 port 47530 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:40.365929 sshd-session[5301]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:40.375878 systemd-logind[1520]: New session 17 of user core. Dec 12 18:41:40.379431 systemd[1]: Started session-17.scope - Session 17 of User core. Dec 12 18:41:41.219513 sshd[5306]: Connection closed by 147.75.109.163 port 47530 Dec 12 18:41:41.220994 sshd-session[5301]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:41.226681 systemd[1]: sshd@16-65.21.248.196:22-147.75.109.163:47530.service: Deactivated successfully. Dec 12 18:41:41.228732 systemd[1]: session-17.scope: Deactivated successfully. Dec 12 18:41:41.230972 systemd-logind[1520]: Session 17 logged out. Waiting for processes to exit. Dec 12 18:41:41.232587 systemd-logind[1520]: Removed session 17. Dec 12 18:41:46.372379 systemd[1]: Started sshd@17-65.21.248.196:22-147.75.109.163:51258.service - OpenSSH per-connection server daemon (147.75.109.163:51258). Dec 12 18:41:47.359833 sshd[5318]: Accepted publickey for core from 147.75.109.163 port 51258 ssh2: RSA SHA256:AJiPuGVqYj3uxkce9uYEDmV2m/PIUxfAon7lAN6EZuk Dec 12 18:41:47.361651 sshd-session[5318]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:41:47.370101 systemd-logind[1520]: New session 18 of user core. Dec 12 18:41:47.375552 systemd[1]: Started session-18.scope - Session 18 of User core. Dec 12 18:41:48.133743 sshd[5321]: Connection closed by 147.75.109.163 port 51258 Dec 12 18:41:48.135707 sshd-session[5318]: pam_unix(sshd:session): session closed for user core Dec 12 18:41:48.138968 systemd[1]: sshd@17-65.21.248.196:22-147.75.109.163:51258.service: Deactivated successfully. Dec 12 18:41:48.140808 systemd[1]: session-18.scope: Deactivated successfully. Dec 12 18:41:48.142531 systemd-logind[1520]: Session 18 logged out. Waiting for processes to exit. Dec 12 18:41:48.144365 systemd-logind[1520]: Removed session 18. Dec 12 18:41:49.123736 kubelet[2737]: E1212 18:41:49.123647 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2msts" podUID="583f364c-28f5-44f4-853f-72900703e8a2" Dec 12 18:41:50.121991 kubelet[2737]: E1212 18:41:50.121743 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-4fs8q" podUID="ae09f5bd-f7e2-460b-b94b-eb370fa59677" Dec 12 18:41:50.124748 kubelet[2737]: E1212 18:41:50.124697 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59789c7c4b-wht4c" podUID="986a62cf-f52b-49ec-a518-51f0449c9112" Dec 12 18:41:52.120452 kubelet[2737]: E1212 18:41:52.120320 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7bbfbd7b8d-c5gdp" podUID="53428c7e-77fb-479f-978b-3877d65c8e58" Dec 12 18:41:53.132419 kubelet[2737]: E1212 18:41:53.132357 2737 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-92gml" podUID="3a064600-3d57-4e0d-80c2-a8041300095d"