Nov 23 23:22:55.805261 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Nov 23 23:22:55.805282 kernel: Linux version 6.12.58-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Sun Nov 23 20:53:53 -00 2025 Nov 23 23:22:55.805291 kernel: KASLR enabled Nov 23 23:22:55.805297 kernel: efi: EFI v2.7 by EDK II Nov 23 23:22:55.805302 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb228018 ACPI 2.0=0xdb9b8018 RNG=0xdb9b8a18 MEMRESERVE=0xdb21fd18 Nov 23 23:22:55.805307 kernel: random: crng init done Nov 23 23:22:55.805314 kernel: Kernel is locked down from EFI Secure Boot; see man kernel_lockdown.7 Nov 23 23:22:55.805320 kernel: secureboot: Secure boot enabled Nov 23 23:22:55.805326 kernel: ACPI: Early table checksum verification disabled Nov 23 23:22:55.805333 kernel: ACPI: RSDP 0x00000000DB9B8018 000024 (v02 BOCHS ) Nov 23 23:22:55.805339 kernel: ACPI: XSDT 0x00000000DB9B8F18 000064 (v01 BOCHS BXPC 00000001 01000013) Nov 23 23:22:55.805345 kernel: ACPI: FACP 0x00000000DB9B8B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805350 kernel: ACPI: DSDT 0x00000000DB904018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805356 kernel: ACPI: APIC 0x00000000DB9B8C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805363 kernel: ACPI: PPTT 0x00000000DB9B8098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805370 kernel: ACPI: GTDT 0x00000000DB9B8818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805376 kernel: ACPI: MCFG 0x00000000DB9B8A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805383 kernel: ACPI: SPCR 0x00000000DB9B8918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805389 kernel: ACPI: DBG2 0x00000000DB9B8998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805394 kernel: ACPI: IORT 0x00000000DB9B8198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Nov 23 23:22:55.805400 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Nov 23 23:22:55.805406 kernel: ACPI: Use ACPI SPCR as default console: No Nov 23 23:22:55.805412 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Nov 23 23:22:55.805418 kernel: NODE_DATA(0) allocated [mem 0xdc737a00-0xdc73efff] Nov 23 23:22:55.805424 kernel: Zone ranges: Nov 23 23:22:55.805431 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Nov 23 23:22:55.805437 kernel: DMA32 empty Nov 23 23:22:55.805443 kernel: Normal empty Nov 23 23:22:55.805449 kernel: Device empty Nov 23 23:22:55.805454 kernel: Movable zone start for each node Nov 23 23:22:55.805460 kernel: Early memory node ranges Nov 23 23:22:55.805466 kernel: node 0: [mem 0x0000000040000000-0x00000000dbb4ffff] Nov 23 23:22:55.805472 kernel: node 0: [mem 0x00000000dbb50000-0x00000000dbe7ffff] Nov 23 23:22:55.805478 kernel: node 0: [mem 0x00000000dbe80000-0x00000000dbe9ffff] Nov 23 23:22:55.805539 kernel: node 0: [mem 0x00000000dbea0000-0x00000000dbedffff] Nov 23 23:22:55.805545 kernel: node 0: [mem 0x00000000dbee0000-0x00000000dbf1ffff] Nov 23 23:22:55.805551 kernel: node 0: [mem 0x00000000dbf20000-0x00000000dbf6ffff] Nov 23 23:22:55.805559 kernel: node 0: [mem 0x00000000dbf70000-0x00000000dcbfffff] Nov 23 23:22:55.805565 kernel: node 0: [mem 0x00000000dcc00000-0x00000000dcfdffff] Nov 23 23:22:55.805571 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Nov 23 23:22:55.805580 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Nov 23 23:22:55.805586 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Nov 23 23:22:55.805592 kernel: cma: Reserved 16 MiB at 0x00000000d7a00000 on node -1 Nov 23 23:22:55.805599 kernel: psci: probing for conduit method from ACPI. Nov 23 23:22:55.805607 kernel: psci: PSCIv1.1 detected in firmware. Nov 23 23:22:55.805613 kernel: psci: Using standard PSCI v0.2 function IDs Nov 23 23:22:55.805620 kernel: psci: Trusted OS migration not required Nov 23 23:22:55.805626 kernel: psci: SMC Calling Convention v1.1 Nov 23 23:22:55.805632 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Nov 23 23:22:55.805639 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Nov 23 23:22:55.805645 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Nov 23 23:22:55.805652 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Nov 23 23:22:55.805658 kernel: Detected PIPT I-cache on CPU0 Nov 23 23:22:55.805665 kernel: CPU features: detected: GIC system register CPU interface Nov 23 23:22:55.805672 kernel: CPU features: detected: Spectre-v4 Nov 23 23:22:55.805678 kernel: CPU features: detected: Spectre-BHB Nov 23 23:22:55.805685 kernel: CPU features: kernel page table isolation forced ON by KASLR Nov 23 23:22:55.805691 kernel: CPU features: detected: Kernel page table isolation (KPTI) Nov 23 23:22:55.805697 kernel: CPU features: detected: ARM erratum 1418040 Nov 23 23:22:55.805704 kernel: CPU features: detected: SSBS not fully self-synchronizing Nov 23 23:22:55.805710 kernel: alternatives: applying boot alternatives Nov 23 23:22:55.805718 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=4db094b704dd398addf25219e01d6d8f197b31dbf6377199102cc61dad0e4bb2 Nov 23 23:22:55.805724 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Nov 23 23:22:55.805731 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 23 23:22:55.805739 kernel: Fallback order for Node 0: 0 Nov 23 23:22:55.805745 kernel: Built 1 zonelists, mobility grouping on. Total pages: 643072 Nov 23 23:22:55.805752 kernel: Policy zone: DMA Nov 23 23:22:55.805758 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Nov 23 23:22:55.805764 kernel: software IO TLB: SWIOTLB bounce buffer size adjusted to 2MB Nov 23 23:22:55.805770 kernel: software IO TLB: area num 4. Nov 23 23:22:55.805777 kernel: software IO TLB: SWIOTLB bounce buffer size roundup to 4MB Nov 23 23:22:55.805783 kernel: software IO TLB: mapped [mem 0x00000000db504000-0x00000000db904000] (4MB) Nov 23 23:22:55.805790 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Nov 23 23:22:55.805796 kernel: rcu: Preemptible hierarchical RCU implementation. Nov 23 23:22:55.805803 kernel: rcu: RCU event tracing is enabled. Nov 23 23:22:55.805809 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Nov 23 23:22:55.805817 kernel: Trampoline variant of Tasks RCU enabled. Nov 23 23:22:55.805824 kernel: Tracing variant of Tasks RCU enabled. Nov 23 23:22:55.805831 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Nov 23 23:22:55.805837 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Nov 23 23:22:55.805843 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 23 23:22:55.805850 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 23 23:22:55.805856 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Nov 23 23:22:55.805863 kernel: GICv3: 256 SPIs implemented Nov 23 23:22:55.805869 kernel: GICv3: 0 Extended SPIs implemented Nov 23 23:22:55.805875 kernel: Root IRQ handler: gic_handle_irq Nov 23 23:22:55.805882 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Nov 23 23:22:55.805888 kernel: GICv3: GICD_CTRL.DS=1, SCR_EL3.FIQ=0 Nov 23 23:22:55.805896 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Nov 23 23:22:55.805902 kernel: ITS [mem 0x08080000-0x0809ffff] Nov 23 23:22:55.805909 kernel: ITS@0x0000000008080000: allocated 8192 Devices @40110000 (indirect, esz 8, psz 64K, shr 1) Nov 23 23:22:55.805915 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @40120000 (flat, esz 8, psz 64K, shr 1) Nov 23 23:22:55.805922 kernel: GICv3: using LPI property table @0x0000000040130000 Nov 23 23:22:55.805928 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040140000 Nov 23 23:22:55.805935 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Nov 23 23:22:55.805941 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Nov 23 23:22:55.805948 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Nov 23 23:22:55.805955 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Nov 23 23:22:55.805961 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Nov 23 23:22:55.805969 kernel: arm-pv: using stolen time PV Nov 23 23:22:55.805976 kernel: Console: colour dummy device 80x25 Nov 23 23:22:55.805982 kernel: ACPI: Core revision 20240827 Nov 23 23:22:55.805989 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Nov 23 23:22:55.805996 kernel: pid_max: default: 32768 minimum: 301 Nov 23 23:22:55.806002 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Nov 23 23:22:55.806009 kernel: landlock: Up and running. Nov 23 23:22:55.806016 kernel: SELinux: Initializing. Nov 23 23:22:55.806022 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 23 23:22:55.806030 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 23 23:22:55.806037 kernel: rcu: Hierarchical SRCU implementation. Nov 23 23:22:55.806044 kernel: rcu: Max phase no-delay instances is 400. Nov 23 23:22:55.806050 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Nov 23 23:22:55.806057 kernel: Remapping and enabling EFI services. Nov 23 23:22:55.806063 kernel: smp: Bringing up secondary CPUs ... Nov 23 23:22:55.806074 kernel: Detected PIPT I-cache on CPU1 Nov 23 23:22:55.806081 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Nov 23 23:22:55.806088 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040150000 Nov 23 23:22:55.806096 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Nov 23 23:22:55.806118 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Nov 23 23:22:55.806130 kernel: Detected PIPT I-cache on CPU2 Nov 23 23:22:55.806140 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Nov 23 23:22:55.806147 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040160000 Nov 23 23:22:55.806154 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Nov 23 23:22:55.806161 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Nov 23 23:22:55.806168 kernel: Detected PIPT I-cache on CPU3 Nov 23 23:22:55.806177 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Nov 23 23:22:55.806184 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040170000 Nov 23 23:22:55.806191 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Nov 23 23:22:55.806197 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Nov 23 23:22:55.806204 kernel: smp: Brought up 1 node, 4 CPUs Nov 23 23:22:55.806211 kernel: SMP: Total of 4 processors activated. Nov 23 23:22:55.806218 kernel: CPU: All CPU(s) started at EL1 Nov 23 23:22:55.806225 kernel: CPU features: detected: 32-bit EL0 Support Nov 23 23:22:55.806232 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Nov 23 23:22:55.806239 kernel: CPU features: detected: Common not Private translations Nov 23 23:22:55.806247 kernel: CPU features: detected: CRC32 instructions Nov 23 23:22:55.806254 kernel: CPU features: detected: Enhanced Virtualization Traps Nov 23 23:22:55.806261 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Nov 23 23:22:55.806268 kernel: CPU features: detected: LSE atomic instructions Nov 23 23:22:55.806275 kernel: CPU features: detected: Privileged Access Never Nov 23 23:22:55.806282 kernel: CPU features: detected: RAS Extension Support Nov 23 23:22:55.806289 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Nov 23 23:22:55.806295 kernel: alternatives: applying system-wide alternatives Nov 23 23:22:55.806302 kernel: CPU features: detected: Hardware dirty bit management on CPU0-3 Nov 23 23:22:55.806311 kernel: Memory: 2421668K/2572288K available (11200K kernel code, 2456K rwdata, 9084K rodata, 39552K init, 1038K bss, 128284K reserved, 16384K cma-reserved) Nov 23 23:22:55.806318 kernel: devtmpfs: initialized Nov 23 23:22:55.806325 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Nov 23 23:22:55.806332 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Nov 23 23:22:55.806339 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Nov 23 23:22:55.806346 kernel: 0 pages in range for non-PLT usage Nov 23 23:22:55.806353 kernel: 508400 pages in range for PLT usage Nov 23 23:22:55.806359 kernel: pinctrl core: initialized pinctrl subsystem Nov 23 23:22:55.806366 kernel: SMBIOS 3.0.0 present. Nov 23 23:22:55.806375 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 Nov 23 23:22:55.806382 kernel: DMI: Memory slots populated: 1/1 Nov 23 23:22:55.806388 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Nov 23 23:22:55.806395 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Nov 23 23:22:55.806402 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Nov 23 23:22:55.806409 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Nov 23 23:22:55.806416 kernel: audit: initializing netlink subsys (disabled) Nov 23 23:22:55.806423 kernel: audit: type=2000 audit(0.023:1): state=initialized audit_enabled=0 res=1 Nov 23 23:22:55.806430 kernel: thermal_sys: Registered thermal governor 'step_wise' Nov 23 23:22:55.806438 kernel: cpuidle: using governor menu Nov 23 23:22:55.806445 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Nov 23 23:22:55.806452 kernel: ASID allocator initialised with 32768 entries Nov 23 23:22:55.806459 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Nov 23 23:22:55.806466 kernel: Serial: AMBA PL011 UART driver Nov 23 23:22:55.806473 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Nov 23 23:22:55.806487 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Nov 23 23:22:55.806495 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Nov 23 23:22:55.806512 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Nov 23 23:22:55.806521 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Nov 23 23:22:55.806528 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Nov 23 23:22:55.806535 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Nov 23 23:22:55.806542 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Nov 23 23:22:55.806549 kernel: ACPI: Added _OSI(Module Device) Nov 23 23:22:55.806556 kernel: ACPI: Added _OSI(Processor Device) Nov 23 23:22:55.806562 kernel: ACPI: Added _OSI(Processor Aggregator Device) Nov 23 23:22:55.806569 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Nov 23 23:22:55.806576 kernel: ACPI: Interpreter enabled Nov 23 23:22:55.806584 kernel: ACPI: Using GIC for interrupt routing Nov 23 23:22:55.806592 kernel: ACPI: MCFG table detected, 1 entries Nov 23 23:22:55.806598 kernel: ACPI: CPU0 has been hot-added Nov 23 23:22:55.806605 kernel: ACPI: CPU1 has been hot-added Nov 23 23:22:55.806612 kernel: ACPI: CPU2 has been hot-added Nov 23 23:22:55.806619 kernel: ACPI: CPU3 has been hot-added Nov 23 23:22:55.806626 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Nov 23 23:22:55.806633 kernel: printk: legacy console [ttyAMA0] enabled Nov 23 23:22:55.806640 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Nov 23 23:22:55.806783 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Nov 23 23:22:55.806850 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Nov 23 23:22:55.806911 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Nov 23 23:22:55.806969 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Nov 23 23:22:55.807029 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Nov 23 23:22:55.807038 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Nov 23 23:22:55.807046 kernel: PCI host bridge to bus 0000:00 Nov 23 23:22:55.807151 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Nov 23 23:22:55.807210 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Nov 23 23:22:55.807294 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Nov 23 23:22:55.807348 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Nov 23 23:22:55.807426 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 conventional PCI endpoint Nov 23 23:22:55.807518 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Nov 23 23:22:55.807587 kernel: pci 0000:00:01.0: BAR 0 [io 0x0000-0x001f] Nov 23 23:22:55.807647 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff] Nov 23 23:22:55.807706 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref] Nov 23 23:22:55.807764 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref]: assigned Nov 23 23:22:55.807824 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff]: assigned Nov 23 23:22:55.807882 kernel: pci 0000:00:01.0: BAR 0 [io 0x1000-0x101f]: assigned Nov 23 23:22:55.807934 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Nov 23 23:22:55.807987 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Nov 23 23:22:55.808039 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Nov 23 23:22:55.808047 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Nov 23 23:22:55.808055 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Nov 23 23:22:55.808062 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Nov 23 23:22:55.808069 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Nov 23 23:22:55.808075 kernel: iommu: Default domain type: Translated Nov 23 23:22:55.808082 kernel: iommu: DMA domain TLB invalidation policy: strict mode Nov 23 23:22:55.808091 kernel: efivars: Registered efivars operations Nov 23 23:22:55.808111 kernel: vgaarb: loaded Nov 23 23:22:55.808119 kernel: clocksource: Switched to clocksource arch_sys_counter Nov 23 23:22:55.808126 kernel: VFS: Disk quotas dquot_6.6.0 Nov 23 23:22:55.808134 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Nov 23 23:22:55.808141 kernel: pnp: PnP ACPI init Nov 23 23:22:55.808212 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Nov 23 23:22:55.808222 kernel: pnp: PnP ACPI: found 1 devices Nov 23 23:22:55.808229 kernel: NET: Registered PF_INET protocol family Nov 23 23:22:55.808238 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 23 23:22:55.808246 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Nov 23 23:22:55.808253 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Nov 23 23:22:55.808260 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Nov 23 23:22:55.808267 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Nov 23 23:22:55.808274 kernel: TCP: Hash tables configured (established 32768 bind 32768) Nov 23 23:22:55.808282 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 23 23:22:55.808289 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 23 23:22:55.808298 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Nov 23 23:22:55.808305 kernel: PCI: CLS 0 bytes, default 64 Nov 23 23:22:55.808312 kernel: kvm [1]: HYP mode not available Nov 23 23:22:55.808319 kernel: Initialise system trusted keyrings Nov 23 23:22:55.808326 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Nov 23 23:22:55.808333 kernel: Key type asymmetric registered Nov 23 23:22:55.808339 kernel: Asymmetric key parser 'x509' registered Nov 23 23:22:55.808346 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Nov 23 23:22:55.808353 kernel: io scheduler mq-deadline registered Nov 23 23:22:55.808360 kernel: io scheduler kyber registered Nov 23 23:22:55.808369 kernel: io scheduler bfq registered Nov 23 23:22:55.808376 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Nov 23 23:22:55.808383 kernel: ACPI: button: Power Button [PWRB] Nov 23 23:22:55.808390 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Nov 23 23:22:55.808448 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Nov 23 23:22:55.808457 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Nov 23 23:22:55.808464 kernel: thunder_xcv, ver 1.0 Nov 23 23:22:55.808471 kernel: thunder_bgx, ver 1.0 Nov 23 23:22:55.808478 kernel: nicpf, ver 1.0 Nov 23 23:22:55.808509 kernel: nicvf, ver 1.0 Nov 23 23:22:55.808584 kernel: rtc-efi rtc-efi.0: registered as rtc0 Nov 23 23:22:55.808639 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-11-23T23:22:55 UTC (1763940175) Nov 23 23:22:55.808648 kernel: hid: raw HID events driver (C) Jiri Kosina Nov 23 23:22:55.808656 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Nov 23 23:22:55.808663 kernel: watchdog: NMI not fully supported Nov 23 23:22:55.808670 kernel: watchdog: Hard watchdog permanently disabled Nov 23 23:22:55.808677 kernel: NET: Registered PF_INET6 protocol family Nov 23 23:22:55.808687 kernel: Segment Routing with IPv6 Nov 23 23:22:55.808694 kernel: In-situ OAM (IOAM) with IPv6 Nov 23 23:22:55.808701 kernel: NET: Registered PF_PACKET protocol family Nov 23 23:22:55.808708 kernel: Key type dns_resolver registered Nov 23 23:22:55.808715 kernel: registered taskstats version 1 Nov 23 23:22:55.808721 kernel: Loading compiled-in X.509 certificates Nov 23 23:22:55.808728 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.58-flatcar: 00c36da29593053a7da9cd3c5945ae69451ce339' Nov 23 23:22:55.808735 kernel: Demotion targets for Node 0: null Nov 23 23:22:55.808742 kernel: Key type .fscrypt registered Nov 23 23:22:55.808750 kernel: Key type fscrypt-provisioning registered Nov 23 23:22:55.808758 kernel: ima: No TPM chip found, activating TPM-bypass! Nov 23 23:22:55.808765 kernel: ima: Allocated hash algorithm: sha1 Nov 23 23:22:55.808772 kernel: ima: No architecture policies found Nov 23 23:22:55.808779 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Nov 23 23:22:55.808786 kernel: clk: Disabling unused clocks Nov 23 23:22:55.808793 kernel: PM: genpd: Disabling unused power domains Nov 23 23:22:55.808800 kernel: Warning: unable to open an initial console. Nov 23 23:22:55.808808 kernel: Freeing unused kernel memory: 39552K Nov 23 23:22:55.808815 kernel: Run /init as init process Nov 23 23:22:55.808822 kernel: with arguments: Nov 23 23:22:55.808829 kernel: /init Nov 23 23:22:55.808835 kernel: with environment: Nov 23 23:22:55.808842 kernel: HOME=/ Nov 23 23:22:55.808849 kernel: TERM=linux Nov 23 23:22:55.808857 systemd[1]: Successfully made /usr/ read-only. Nov 23 23:22:55.808868 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Nov 23 23:22:55.808877 systemd[1]: Detected virtualization kvm. Nov 23 23:22:55.808885 systemd[1]: Detected architecture arm64. Nov 23 23:22:55.808892 systemd[1]: Running in initrd. Nov 23 23:22:55.808900 systemd[1]: No hostname configured, using default hostname. Nov 23 23:22:55.808907 systemd[1]: Hostname set to . Nov 23 23:22:55.808914 systemd[1]: Initializing machine ID from VM UUID. Nov 23 23:22:55.808922 systemd[1]: Queued start job for default target initrd.target. Nov 23 23:22:55.808931 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 23 23:22:55.808939 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 23 23:22:55.808947 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Nov 23 23:22:55.808955 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 23 23:22:55.808962 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Nov 23 23:22:55.808971 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Nov 23 23:22:55.808979 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Nov 23 23:22:55.808988 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Nov 23 23:22:55.808996 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 23 23:22:55.809003 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 23 23:22:55.809011 systemd[1]: Reached target paths.target - Path Units. Nov 23 23:22:55.809018 systemd[1]: Reached target slices.target - Slice Units. Nov 23 23:22:55.809026 systemd[1]: Reached target swap.target - Swaps. Nov 23 23:22:55.809033 systemd[1]: Reached target timers.target - Timer Units. Nov 23 23:22:55.809040 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Nov 23 23:22:55.809048 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 23 23:22:55.809057 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Nov 23 23:22:55.809065 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Nov 23 23:22:55.809073 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 23 23:22:55.809081 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 23 23:22:55.809088 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 23 23:22:55.809096 systemd[1]: Reached target sockets.target - Socket Units. Nov 23 23:22:55.809111 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Nov 23 23:22:55.809120 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 23 23:22:55.809130 systemd[1]: Finished network-cleanup.service - Network Cleanup. Nov 23 23:22:55.809138 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Nov 23 23:22:55.809146 systemd[1]: Starting systemd-fsck-usr.service... Nov 23 23:22:55.809154 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 23 23:22:55.809162 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 23 23:22:55.809169 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 23 23:22:55.809177 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Nov 23 23:22:55.809187 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 23 23:22:55.809195 systemd[1]: Finished systemd-fsck-usr.service. Nov 23 23:22:55.809220 systemd-journald[245]: Collecting audit messages is disabled. Nov 23 23:22:55.809241 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 23 23:22:55.809250 systemd-journald[245]: Journal started Nov 23 23:22:55.809268 systemd-journald[245]: Runtime Journal (/run/log/journal/13a223e77f6d4e81992a311d1f58a6fd) is 6M, max 48.5M, 42.4M free. Nov 23 23:22:55.801327 systemd-modules-load[247]: Inserted module 'overlay' Nov 23 23:22:55.811235 systemd[1]: Started systemd-journald.service - Journal Service. Nov 23 23:22:55.818643 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 23 23:22:55.821588 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 23 23:22:55.821611 kernel: Bridge firewalling registered Nov 23 23:22:55.821985 systemd-modules-load[247]: Inserted module 'br_netfilter' Nov 23 23:22:55.822914 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 23 23:22:55.824864 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 23 23:22:55.827988 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 23 23:22:55.842660 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 23 23:22:55.848041 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 23 23:22:55.849831 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 23 23:22:55.849884 systemd-tmpfiles[264]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Nov 23 23:22:55.853540 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 23 23:22:55.864916 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 23 23:22:55.866403 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 23 23:22:55.868700 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 23 23:22:55.872209 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Nov 23 23:22:55.874855 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 23 23:22:55.902271 dracut-cmdline[288]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=4db094b704dd398addf25219e01d6d8f197b31dbf6377199102cc61dad0e4bb2 Nov 23 23:22:55.917176 systemd-resolved[289]: Positive Trust Anchors: Nov 23 23:22:55.917198 systemd-resolved[289]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 23 23:22:55.917239 systemd-resolved[289]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 23 23:22:55.922156 systemd-resolved[289]: Defaulting to hostname 'linux'. Nov 23 23:22:55.923264 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 23 23:22:55.928241 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 23 23:22:55.981515 kernel: SCSI subsystem initialized Nov 23 23:22:55.986506 kernel: Loading iSCSI transport class v2.0-870. Nov 23 23:22:55.994513 kernel: iscsi: registered transport (tcp) Nov 23 23:22:56.007518 kernel: iscsi: registered transport (qla4xxx) Nov 23 23:22:56.007545 kernel: QLogic iSCSI HBA Driver Nov 23 23:22:56.024524 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 23 23:22:56.048357 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 23 23:22:56.050847 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 23 23:22:56.099095 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Nov 23 23:22:56.103625 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Nov 23 23:22:56.160515 kernel: raid6: neonx8 gen() 15785 MB/s Nov 23 23:22:56.177507 kernel: raid6: neonx4 gen() 15744 MB/s Nov 23 23:22:56.194510 kernel: raid6: neonx2 gen() 13034 MB/s Nov 23 23:22:56.211509 kernel: raid6: neonx1 gen() 10357 MB/s Nov 23 23:22:56.228510 kernel: raid6: int64x8 gen() 6792 MB/s Nov 23 23:22:56.245508 kernel: raid6: int64x4 gen() 7338 MB/s Nov 23 23:22:56.262510 kernel: raid6: int64x2 gen() 5988 MB/s Nov 23 23:22:56.279751 kernel: raid6: int64x1 gen() 5012 MB/s Nov 23 23:22:56.279771 kernel: raid6: using algorithm neonx8 gen() 15785 MB/s Nov 23 23:22:56.297644 kernel: raid6: .... xor() 11846 MB/s, rmw enabled Nov 23 23:22:56.297659 kernel: raid6: using neon recovery algorithm Nov 23 23:22:56.303975 kernel: xor: measuring software checksum speed Nov 23 23:22:56.303997 kernel: 8regs : 21618 MB/sec Nov 23 23:22:56.304008 kernel: 32regs : 21676 MB/sec Nov 23 23:22:56.304625 kernel: arm64_neon : 28003 MB/sec Nov 23 23:22:56.304639 kernel: xor: using function: arm64_neon (28003 MB/sec) Nov 23 23:22:56.357512 kernel: Btrfs loaded, zoned=no, fsverity=no Nov 23 23:22:56.364557 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Nov 23 23:22:56.367083 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 23 23:22:56.392804 systemd-udevd[497]: Using default interface naming scheme 'v255'. Nov 23 23:22:56.396840 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 23 23:22:56.399600 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Nov 23 23:22:56.435181 dracut-pre-trigger[507]: rd.md=0: removing MD RAID activation Nov 23 23:22:56.478535 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Nov 23 23:22:56.480961 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 23 23:22:56.538053 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 23 23:22:56.540830 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Nov 23 23:22:56.588050 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Nov 23 23:22:56.588224 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Nov 23 23:22:56.594793 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 23 23:22:56.605020 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Nov 23 23:22:56.605043 kernel: GPT:9289727 != 19775487 Nov 23 23:22:56.594913 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 23 23:22:56.608178 kernel: GPT:Alternate GPT header not at the end of the disk. Nov 23 23:22:56.608206 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Nov 23 23:22:56.611015 kernel: GPT:9289727 != 19775487 Nov 23 23:22:56.611087 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 23 23:22:56.614160 kernel: GPT: Use GNU Parted to correct GPT errors. Nov 23 23:22:56.614185 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 23 23:22:56.640541 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Nov 23 23:22:56.643328 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 23 23:22:56.645893 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Nov 23 23:22:56.658656 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Nov 23 23:22:56.664949 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Nov 23 23:22:56.666221 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Nov 23 23:22:56.675199 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Nov 23 23:22:56.676515 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Nov 23 23:22:56.678792 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 23 23:22:56.680990 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 23 23:22:56.683786 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Nov 23 23:22:56.685663 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Nov 23 23:22:56.702801 disk-uuid[591]: Primary Header is updated. Nov 23 23:22:56.702801 disk-uuid[591]: Secondary Entries is updated. Nov 23 23:22:56.702801 disk-uuid[591]: Secondary Header is updated. Nov 23 23:22:56.707511 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 23 23:22:56.707854 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Nov 23 23:22:57.719508 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 23 23:22:57.720585 disk-uuid[595]: The operation has completed successfully. Nov 23 23:22:57.745115 systemd[1]: disk-uuid.service: Deactivated successfully. Nov 23 23:22:57.745215 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Nov 23 23:22:57.769311 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Nov 23 23:22:57.791645 sh[610]: Success Nov 23 23:22:57.804543 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Nov 23 23:22:57.805133 kernel: device-mapper: uevent: version 1.0.3 Nov 23 23:22:57.805147 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Nov 23 23:22:57.813504 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Nov 23 23:22:57.841254 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Nov 23 23:22:57.843332 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Nov 23 23:22:57.857003 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Nov 23 23:22:57.864232 kernel: BTRFS: device fsid 5fd06d80-8dd4-4ca0-aa0c-93ddab5f4498 devid 1 transid 38 /dev/mapper/usr (253:0) scanned by mount (622) Nov 23 23:22:57.864266 kernel: BTRFS info (device dm-0): first mount of filesystem 5fd06d80-8dd4-4ca0-aa0c-93ddab5f4498 Nov 23 23:22:57.864277 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Nov 23 23:22:57.869631 kernel: BTRFS info (device dm-0): disabling log replay at mount time Nov 23 23:22:57.869668 kernel: BTRFS info (device dm-0): enabling free space tree Nov 23 23:22:57.871132 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Nov 23 23:22:57.872559 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Nov 23 23:22:57.874003 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Nov 23 23:22:57.874810 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Nov 23 23:22:57.876527 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Nov 23 23:22:57.899507 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (652) Nov 23 23:22:57.899558 kernel: BTRFS info (device vda6): first mount of filesystem fbc9a6bc-8b9c-4847-949c-e8c4f3bf01b3 Nov 23 23:22:57.901605 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Nov 23 23:22:57.904228 kernel: BTRFS info (device vda6): turning on async discard Nov 23 23:22:57.904340 kernel: BTRFS info (device vda6): enabling free space tree Nov 23 23:22:57.908503 kernel: BTRFS info (device vda6): last unmount of filesystem fbc9a6bc-8b9c-4847-949c-e8c4f3bf01b3 Nov 23 23:22:57.909469 systemd[1]: Finished ignition-setup.service - Ignition (setup). Nov 23 23:22:57.912416 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Nov 23 23:22:57.984147 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 23 23:22:57.987930 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 23 23:22:58.017767 ignition[701]: Ignition 2.22.0 Nov 23 23:22:58.017786 ignition[701]: Stage: fetch-offline Nov 23 23:22:58.017823 ignition[701]: no configs at "/usr/lib/ignition/base.d" Nov 23 23:22:58.017830 ignition[701]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 23 23:22:58.017913 ignition[701]: parsed url from cmdline: "" Nov 23 23:22:58.017916 ignition[701]: no config URL provided Nov 23 23:22:58.017921 ignition[701]: reading system config file "/usr/lib/ignition/user.ign" Nov 23 23:22:58.017928 ignition[701]: no config at "/usr/lib/ignition/user.ign" Nov 23 23:22:58.017947 ignition[701]: op(1): [started] loading QEMU firmware config module Nov 23 23:22:58.017952 ignition[701]: op(1): executing: "modprobe" "qemu_fw_cfg" Nov 23 23:22:58.023832 ignition[701]: op(1): [finished] loading QEMU firmware config module Nov 23 23:22:58.028470 systemd-networkd[806]: lo: Link UP Nov 23 23:22:58.028497 systemd-networkd[806]: lo: Gained carrier Nov 23 23:22:58.029150 systemd-networkd[806]: Enumeration completed Nov 23 23:22:58.029250 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 23 23:22:58.029531 systemd-networkd[806]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 23 23:22:58.029535 systemd-networkd[806]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 23 23:22:58.030909 systemd-networkd[806]: eth0: Link UP Nov 23 23:22:58.030993 systemd-networkd[806]: eth0: Gained carrier Nov 23 23:22:58.031002 systemd-networkd[806]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 23 23:22:58.031042 systemd[1]: Reached target network.target - Network. Nov 23 23:22:58.048553 systemd-networkd[806]: eth0: DHCPv4 address 10.0.0.134/16, gateway 10.0.0.1 acquired from 10.0.0.1 Nov 23 23:22:58.082400 ignition[701]: parsing config with SHA512: 35b340b01b2d89a297a9688e2565c43253495618bf55c2adbbd45560a99ee0f7617b4bfd215b88abb911136af10f118912543db09796567ae081ae1acf88facb Nov 23 23:22:58.088444 unknown[701]: fetched base config from "system" Nov 23 23:22:58.088460 unknown[701]: fetched user config from "qemu" Nov 23 23:22:58.088928 ignition[701]: fetch-offline: fetch-offline passed Nov 23 23:22:58.088986 ignition[701]: Ignition finished successfully Nov 23 23:22:58.093249 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Nov 23 23:22:58.095659 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Nov 23 23:22:58.096406 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Nov 23 23:22:58.135691 ignition[814]: Ignition 2.22.0 Nov 23 23:22:58.135710 ignition[814]: Stage: kargs Nov 23 23:22:58.135844 ignition[814]: no configs at "/usr/lib/ignition/base.d" Nov 23 23:22:58.135853 ignition[814]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 23 23:22:58.136595 ignition[814]: kargs: kargs passed Nov 23 23:22:58.139721 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Nov 23 23:22:58.136640 ignition[814]: Ignition finished successfully Nov 23 23:22:58.141731 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Nov 23 23:22:58.178540 ignition[822]: Ignition 2.22.0 Nov 23 23:22:58.178556 ignition[822]: Stage: disks Nov 23 23:22:58.178688 ignition[822]: no configs at "/usr/lib/ignition/base.d" Nov 23 23:22:58.178697 ignition[822]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 23 23:22:58.179441 ignition[822]: disks: disks passed Nov 23 23:22:58.179498 ignition[822]: Ignition finished successfully Nov 23 23:22:58.184273 systemd[1]: Finished ignition-disks.service - Ignition (disks). Nov 23 23:22:58.185868 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Nov 23 23:22:58.187663 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Nov 23 23:22:58.189671 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 23 23:22:58.191568 systemd[1]: Reached target sysinit.target - System Initialization. Nov 23 23:22:58.193338 systemd[1]: Reached target basic.target - Basic System. Nov 23 23:22:58.195929 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Nov 23 23:22:58.231957 systemd-fsck[832]: ROOT: clean, 15/553520 files, 52789/553472 blocks Nov 23 23:22:58.390856 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Nov 23 23:22:58.393209 systemd[1]: Mounting sysroot.mount - /sysroot... Nov 23 23:22:58.481517 kernel: EXT4-fs (vda9): mounted filesystem fa3f8731-d4e3-4e51-b6db-fa404206cf07 r/w with ordered data mode. Quota mode: none. Nov 23 23:22:58.482384 systemd[1]: Mounted sysroot.mount - /sysroot. Nov 23 23:22:58.483719 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Nov 23 23:22:58.486144 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 23 23:22:58.487981 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Nov 23 23:22:58.489125 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Nov 23 23:22:58.489167 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Nov 23 23:22:58.489189 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Nov 23 23:22:58.496460 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Nov 23 23:22:58.499778 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Nov 23 23:22:58.502536 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (842) Nov 23 23:22:58.504637 kernel: BTRFS info (device vda6): first mount of filesystem fbc9a6bc-8b9c-4847-949c-e8c4f3bf01b3 Nov 23 23:22:58.504658 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Nov 23 23:22:58.507510 kernel: BTRFS info (device vda6): turning on async discard Nov 23 23:22:58.507548 kernel: BTRFS info (device vda6): enabling free space tree Nov 23 23:22:58.509334 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 23 23:22:58.536706 initrd-setup-root[866]: cut: /sysroot/etc/passwd: No such file or directory Nov 23 23:22:58.541421 initrd-setup-root[873]: cut: /sysroot/etc/group: No such file or directory Nov 23 23:22:58.545435 initrd-setup-root[880]: cut: /sysroot/etc/shadow: No such file or directory Nov 23 23:22:58.549108 initrd-setup-root[887]: cut: /sysroot/etc/gshadow: No such file or directory Nov 23 23:22:58.614036 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Nov 23 23:22:58.616169 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Nov 23 23:22:58.617861 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Nov 23 23:22:58.636506 kernel: BTRFS info (device vda6): last unmount of filesystem fbc9a6bc-8b9c-4847-949c-e8c4f3bf01b3 Nov 23 23:22:58.652633 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Nov 23 23:22:58.666859 ignition[955]: INFO : Ignition 2.22.0 Nov 23 23:22:58.666859 ignition[955]: INFO : Stage: mount Nov 23 23:22:58.668510 ignition[955]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 23 23:22:58.668510 ignition[955]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 23 23:22:58.668510 ignition[955]: INFO : mount: mount passed Nov 23 23:22:58.668510 ignition[955]: INFO : Ignition finished successfully Nov 23 23:22:58.669300 systemd[1]: Finished ignition-mount.service - Ignition (mount). Nov 23 23:22:58.671390 systemd[1]: Starting ignition-files.service - Ignition (files)... Nov 23 23:22:58.862305 systemd[1]: sysroot-oem.mount: Deactivated successfully. Nov 23 23:22:58.866681 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 23 23:22:58.884535 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (970) Nov 23 23:22:58.884583 kernel: BTRFS info (device vda6): first mount of filesystem fbc9a6bc-8b9c-4847-949c-e8c4f3bf01b3 Nov 23 23:22:58.884594 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Nov 23 23:22:58.888513 kernel: BTRFS info (device vda6): turning on async discard Nov 23 23:22:58.888543 kernel: BTRFS info (device vda6): enabling free space tree Nov 23 23:22:58.889928 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 23 23:22:58.927468 ignition[987]: INFO : Ignition 2.22.0 Nov 23 23:22:58.927468 ignition[987]: INFO : Stage: files Nov 23 23:22:58.929254 ignition[987]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 23 23:22:58.929254 ignition[987]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 23 23:22:58.929254 ignition[987]: DEBUG : files: compiled without relabeling support, skipping Nov 23 23:22:58.932922 ignition[987]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Nov 23 23:22:58.932922 ignition[987]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Nov 23 23:22:58.935812 ignition[987]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Nov 23 23:22:58.937262 ignition[987]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Nov 23 23:22:58.937262 ignition[987]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Nov 23 23:22:58.936311 unknown[987]: wrote ssh authorized keys file for user: core Nov 23 23:22:58.941386 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Nov 23 23:22:58.941386 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-arm64.tar.gz: attempt #1 Nov 23 23:22:58.986574 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Nov 23 23:22:59.213506 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Nov 23 23:22:59.213506 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Nov 23 23:22:59.217699 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Nov 23 23:22:59.217699 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Nov 23 23:22:59.217699 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Nov 23 23:22:59.217699 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 23 23:22:59.217699 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 23 23:22:59.217699 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 23 23:22:59.217699 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 23 23:22:59.230537 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Nov 23 23:22:59.230537 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Nov 23 23:22:59.230537 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Nov 23 23:22:59.230537 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Nov 23 23:22:59.230537 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Nov 23 23:22:59.230537 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-arm64.raw: attempt #1 Nov 23 23:22:59.568904 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Nov 23 23:22:59.851119 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Nov 23 23:22:59.853956 ignition[987]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Nov 23 23:22:59.853956 ignition[987]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 23 23:22:59.858299 ignition[987]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 23 23:22:59.858299 ignition[987]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Nov 23 23:22:59.858299 ignition[987]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Nov 23 23:22:59.858299 ignition[987]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Nov 23 23:22:59.858299 ignition[987]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Nov 23 23:22:59.858299 ignition[987]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Nov 23 23:22:59.858299 ignition[987]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Nov 23 23:22:59.875600 systemd-networkd[806]: eth0: Gained IPv6LL Nov 23 23:22:59.877873 ignition[987]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Nov 23 23:22:59.881829 ignition[987]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Nov 23 23:22:59.884069 ignition[987]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Nov 23 23:22:59.884069 ignition[987]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Nov 23 23:22:59.884069 ignition[987]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Nov 23 23:22:59.884069 ignition[987]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Nov 23 23:22:59.884069 ignition[987]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Nov 23 23:22:59.884069 ignition[987]: INFO : files: files passed Nov 23 23:22:59.884069 ignition[987]: INFO : Ignition finished successfully Nov 23 23:22:59.885188 systemd[1]: Finished ignition-files.service - Ignition (files). Nov 23 23:22:59.890421 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Nov 23 23:22:59.893894 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Nov 23 23:22:59.910074 systemd[1]: ignition-quench.service: Deactivated successfully. Nov 23 23:22:59.910423 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Nov 23 23:22:59.914254 initrd-setup-root-after-ignition[1016]: grep: /sysroot/oem/oem-release: No such file or directory Nov 23 23:22:59.915744 initrd-setup-root-after-ignition[1018]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 23 23:22:59.915744 initrd-setup-root-after-ignition[1018]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Nov 23 23:22:59.919224 initrd-setup-root-after-ignition[1022]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 23 23:22:59.920566 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 23 23:22:59.922280 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Nov 23 23:22:59.925352 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Nov 23 23:22:59.957464 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Nov 23 23:22:59.957616 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Nov 23 23:22:59.959884 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Nov 23 23:22:59.961885 systemd[1]: Reached target initrd.target - Initrd Default Target. Nov 23 23:22:59.966594 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Nov 23 23:22:59.967348 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Nov 23 23:22:59.993689 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 23 23:22:59.996233 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Nov 23 23:23:00.022635 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Nov 23 23:23:00.023884 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 23 23:23:00.026086 systemd[1]: Stopped target timers.target - Timer Units. Nov 23 23:23:00.027944 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Nov 23 23:23:00.028064 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 23 23:23:00.030803 systemd[1]: Stopped target initrd.target - Initrd Default Target. Nov 23 23:23:00.032877 systemd[1]: Stopped target basic.target - Basic System. Nov 23 23:23:00.034542 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Nov 23 23:23:00.036382 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Nov 23 23:23:00.038553 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Nov 23 23:23:00.040704 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Nov 23 23:23:00.042702 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Nov 23 23:23:00.044657 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Nov 23 23:23:00.046702 systemd[1]: Stopped target sysinit.target - System Initialization. Nov 23 23:23:00.048689 systemd[1]: Stopped target local-fs.target - Local File Systems. Nov 23 23:23:00.050504 systemd[1]: Stopped target swap.target - Swaps. Nov 23 23:23:00.052157 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Nov 23 23:23:00.052288 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Nov 23 23:23:00.054716 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Nov 23 23:23:00.056817 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 23 23:23:00.058859 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Nov 23 23:23:00.058938 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 23 23:23:00.061127 systemd[1]: dracut-initqueue.service: Deactivated successfully. Nov 23 23:23:00.061239 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Nov 23 23:23:00.064324 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Nov 23 23:23:00.064434 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Nov 23 23:23:00.066524 systemd[1]: Stopped target paths.target - Path Units. Nov 23 23:23:00.068252 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Nov 23 23:23:00.071519 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 23 23:23:00.073664 systemd[1]: Stopped target slices.target - Slice Units. Nov 23 23:23:00.075763 systemd[1]: Stopped target sockets.target - Socket Units. Nov 23 23:23:00.077362 systemd[1]: iscsid.socket: Deactivated successfully. Nov 23 23:23:00.077440 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Nov 23 23:23:00.079156 systemd[1]: iscsiuio.socket: Deactivated successfully. Nov 23 23:23:00.079236 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 23 23:23:00.080808 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Nov 23 23:23:00.080912 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 23 23:23:00.082844 systemd[1]: ignition-files.service: Deactivated successfully. Nov 23 23:23:00.082945 systemd[1]: Stopped ignition-files.service - Ignition (files). Nov 23 23:23:00.085328 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Nov 23 23:23:00.086919 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Nov 23 23:23:00.088344 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Nov 23 23:23:00.088468 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Nov 23 23:23:00.090892 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Nov 23 23:23:00.091011 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Nov 23 23:23:00.096251 systemd[1]: initrd-cleanup.service: Deactivated successfully. Nov 23 23:23:00.104301 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Nov 23 23:23:00.113823 systemd[1]: sysroot-boot.mount: Deactivated successfully. Nov 23 23:23:00.123429 ignition[1043]: INFO : Ignition 2.22.0 Nov 23 23:23:00.123429 ignition[1043]: INFO : Stage: umount Nov 23 23:23:00.126705 ignition[1043]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 23 23:23:00.126705 ignition[1043]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 23 23:23:00.126705 ignition[1043]: INFO : umount: umount passed Nov 23 23:23:00.126705 ignition[1043]: INFO : Ignition finished successfully Nov 23 23:23:00.126854 systemd[1]: ignition-mount.service: Deactivated successfully. Nov 23 23:23:00.126969 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Nov 23 23:23:00.129036 systemd[1]: Stopped target network.target - Network. Nov 23 23:23:00.130912 systemd[1]: ignition-disks.service: Deactivated successfully. Nov 23 23:23:00.130973 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Nov 23 23:23:00.132646 systemd[1]: ignition-kargs.service: Deactivated successfully. Nov 23 23:23:00.132690 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Nov 23 23:23:00.134476 systemd[1]: ignition-setup.service: Deactivated successfully. Nov 23 23:23:00.134546 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Nov 23 23:23:00.136593 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Nov 23 23:23:00.136634 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Nov 23 23:23:00.138703 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Nov 23 23:23:00.140582 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Nov 23 23:23:00.145214 systemd[1]: systemd-resolved.service: Deactivated successfully. Nov 23 23:23:00.145323 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Nov 23 23:23:00.148472 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Nov 23 23:23:00.148781 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Nov 23 23:23:00.148815 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 23 23:23:00.153327 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Nov 23 23:23:00.153539 systemd[1]: systemd-networkd.service: Deactivated successfully. Nov 23 23:23:00.153639 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Nov 23 23:23:00.158211 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Nov 23 23:23:00.158576 systemd[1]: Stopped target network-pre.target - Preparation for Network. Nov 23 23:23:00.160626 systemd[1]: systemd-networkd.socket: Deactivated successfully. Nov 23 23:23:00.160666 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Nov 23 23:23:00.163585 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Nov 23 23:23:00.165062 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Nov 23 23:23:00.165129 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 23 23:23:00.167979 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 23 23:23:00.168086 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 23 23:23:00.171098 systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 23 23:23:00.171139 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Nov 23 23:23:00.173477 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 23 23:23:00.178375 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Nov 23 23:23:00.178794 systemd[1]: sysroot-boot.service: Deactivated successfully. Nov 23 23:23:00.178868 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Nov 23 23:23:00.182574 systemd[1]: initrd-setup-root.service: Deactivated successfully. Nov 23 23:23:00.182632 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Nov 23 23:23:00.193139 systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 23 23:23:00.193285 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 23 23:23:00.195925 systemd[1]: network-cleanup.service: Deactivated successfully. Nov 23 23:23:00.196021 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Nov 23 23:23:00.198045 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Nov 23 23:23:00.198133 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Nov 23 23:23:00.199768 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Nov 23 23:23:00.199803 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Nov 23 23:23:00.201767 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Nov 23 23:23:00.201820 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Nov 23 23:23:00.205116 systemd[1]: dracut-cmdline.service: Deactivated successfully. Nov 23 23:23:00.205166 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Nov 23 23:23:00.208121 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 23 23:23:00.208170 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 23 23:23:00.211314 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Nov 23 23:23:00.212687 systemd[1]: systemd-network-generator.service: Deactivated successfully. Nov 23 23:23:00.212742 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Nov 23 23:23:00.215872 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Nov 23 23:23:00.215917 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 23 23:23:00.219032 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Nov 23 23:23:00.219076 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 23 23:23:00.222365 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Nov 23 23:23:00.222408 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Nov 23 23:23:00.225065 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 23 23:23:00.225122 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 23 23:23:00.228945 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Nov 23 23:23:00.229041 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Nov 23 23:23:00.231610 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Nov 23 23:23:00.234148 systemd[1]: Starting initrd-switch-root.service - Switch Root... Nov 23 23:23:00.248888 systemd[1]: Switching root. Nov 23 23:23:00.284723 systemd-journald[245]: Journal stopped Nov 23 23:23:01.056564 systemd-journald[245]: Received SIGTERM from PID 1 (systemd). Nov 23 23:23:01.056614 kernel: SELinux: policy capability network_peer_controls=1 Nov 23 23:23:01.056627 kernel: SELinux: policy capability open_perms=1 Nov 23 23:23:01.056641 kernel: SELinux: policy capability extended_socket_class=1 Nov 23 23:23:01.056649 kernel: SELinux: policy capability always_check_network=0 Nov 23 23:23:01.056658 kernel: SELinux: policy capability cgroup_seclabel=1 Nov 23 23:23:01.056667 kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 23 23:23:01.056678 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Nov 23 23:23:01.056689 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Nov 23 23:23:01.056698 kernel: SELinux: policy capability userspace_initial_context=0 Nov 23 23:23:01.056707 kernel: audit: type=1403 audit(1763940180.461:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Nov 23 23:23:01.056717 systemd[1]: Successfully loaded SELinux policy in 64.571ms. Nov 23 23:23:01.056737 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 5.286ms. Nov 23 23:23:01.056749 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Nov 23 23:23:01.056761 systemd[1]: Detected virtualization kvm. Nov 23 23:23:01.056772 systemd[1]: Detected architecture arm64. Nov 23 23:23:01.056784 systemd[1]: Detected first boot. Nov 23 23:23:01.056793 systemd[1]: Initializing machine ID from VM UUID. Nov 23 23:23:01.056804 zram_generator::config[1088]: No configuration found. Nov 23 23:23:01.056814 kernel: NET: Registered PF_VSOCK protocol family Nov 23 23:23:01.056823 systemd[1]: Populated /etc with preset unit settings. Nov 23 23:23:01.056840 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Nov 23 23:23:01.056851 systemd[1]: initrd-switch-root.service: Deactivated successfully. Nov 23 23:23:01.056864 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Nov 23 23:23:01.056874 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Nov 23 23:23:01.056884 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Nov 23 23:23:01.056894 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Nov 23 23:23:01.056904 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Nov 23 23:23:01.056915 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Nov 23 23:23:01.056925 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Nov 23 23:23:01.056937 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Nov 23 23:23:01.056947 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Nov 23 23:23:01.056957 systemd[1]: Created slice user.slice - User and Session Slice. Nov 23 23:23:01.056967 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 23 23:23:01.056978 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 23 23:23:01.056988 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Nov 23 23:23:01.056997 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Nov 23 23:23:01.057007 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Nov 23 23:23:01.057018 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 23 23:23:01.057028 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Nov 23 23:23:01.057039 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 23 23:23:01.057049 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 23 23:23:01.057059 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Nov 23 23:23:01.057069 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Nov 23 23:23:01.057086 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Nov 23 23:23:01.057098 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Nov 23 23:23:01.057119 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 23 23:23:01.057132 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 23 23:23:01.057142 systemd[1]: Reached target slices.target - Slice Units. Nov 23 23:23:01.057153 systemd[1]: Reached target swap.target - Swaps. Nov 23 23:23:01.057163 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Nov 23 23:23:01.057174 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Nov 23 23:23:01.057184 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Nov 23 23:23:01.057193 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 23 23:23:01.057204 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 23 23:23:01.057214 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 23 23:23:01.057225 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Nov 23 23:23:01.057235 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Nov 23 23:23:01.057245 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Nov 23 23:23:01.057256 systemd[1]: Mounting media.mount - External Media Directory... Nov 23 23:23:01.057266 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Nov 23 23:23:01.057276 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Nov 23 23:23:01.057286 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Nov 23 23:23:01.057296 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Nov 23 23:23:01.057309 systemd[1]: Reached target machines.target - Containers. Nov 23 23:23:01.057320 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Nov 23 23:23:01.057330 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 23 23:23:01.057340 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 23 23:23:01.057350 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Nov 23 23:23:01.057361 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 23 23:23:01.057370 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 23 23:23:01.057380 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 23 23:23:01.057390 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Nov 23 23:23:01.057401 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 23 23:23:01.057412 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Nov 23 23:23:01.057423 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Nov 23 23:23:01.057433 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Nov 23 23:23:01.057443 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Nov 23 23:23:01.057453 systemd[1]: Stopped systemd-fsck-usr.service. Nov 23 23:23:01.057464 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 23 23:23:01.057473 kernel: fuse: init (API version 7.41) Nov 23 23:23:01.057540 kernel: ACPI: bus type drm_connector registered Nov 23 23:23:01.057552 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 23 23:23:01.057562 kernel: loop: module loaded Nov 23 23:23:01.057571 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 23 23:23:01.057582 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 23 23:23:01.057591 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Nov 23 23:23:01.057601 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Nov 23 23:23:01.057612 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 23 23:23:01.057622 systemd[1]: verity-setup.service: Deactivated successfully. Nov 23 23:23:01.057643 systemd[1]: Stopped verity-setup.service. Nov 23 23:23:01.057685 systemd-journald[1163]: Collecting audit messages is disabled. Nov 23 23:23:01.057708 systemd-journald[1163]: Journal started Nov 23 23:23:01.057729 systemd-journald[1163]: Runtime Journal (/run/log/journal/13a223e77f6d4e81992a311d1f58a6fd) is 6M, max 48.5M, 42.4M free. Nov 23 23:23:00.815961 systemd[1]: Queued start job for default target multi-user.target. Nov 23 23:23:00.838506 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Nov 23 23:23:00.838887 systemd[1]: systemd-journald.service: Deactivated successfully. Nov 23 23:23:01.061438 systemd[1]: Started systemd-journald.service - Journal Service. Nov 23 23:23:01.062163 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Nov 23 23:23:01.063432 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Nov 23 23:23:01.064816 systemd[1]: Mounted media.mount - External Media Directory. Nov 23 23:23:01.065978 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Nov 23 23:23:01.067237 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Nov 23 23:23:01.068526 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Nov 23 23:23:01.069737 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Nov 23 23:23:01.071207 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 23 23:23:01.072794 systemd[1]: modprobe@configfs.service: Deactivated successfully. Nov 23 23:23:01.072955 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Nov 23 23:23:01.074401 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 23 23:23:01.074595 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 23 23:23:01.075988 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 23 23:23:01.076168 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 23 23:23:01.077593 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 23 23:23:01.077750 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 23 23:23:01.079215 systemd[1]: modprobe@fuse.service: Deactivated successfully. Nov 23 23:23:01.079369 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Nov 23 23:23:01.080963 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 23 23:23:01.081145 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 23 23:23:01.082585 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 23 23:23:01.083961 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 23 23:23:01.085516 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Nov 23 23:23:01.087016 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Nov 23 23:23:01.098345 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 23 23:23:01.100699 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Nov 23 23:23:01.102655 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Nov 23 23:23:01.103967 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Nov 23 23:23:01.103994 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 23 23:23:01.105836 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Nov 23 23:23:01.118530 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Nov 23 23:23:01.119782 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 23 23:23:01.121203 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Nov 23 23:23:01.123468 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Nov 23 23:23:01.124869 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 23 23:23:01.125814 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Nov 23 23:23:01.127134 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 23 23:23:01.129258 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 23 23:23:01.131677 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Nov 23 23:23:01.138580 systemd-journald[1163]: Time spent on flushing to /var/log/journal/13a223e77f6d4e81992a311d1f58a6fd is 25.435ms for 881 entries. Nov 23 23:23:01.138580 systemd-journald[1163]: System Journal (/var/log/journal/13a223e77f6d4e81992a311d1f58a6fd) is 8M, max 195.6M, 187.6M free. Nov 23 23:23:01.173730 systemd-journald[1163]: Received client request to flush runtime journal. Nov 23 23:23:01.173771 kernel: loop0: detected capacity change from 0 to 207008 Nov 23 23:23:01.173865 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Nov 23 23:23:01.136645 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 23 23:23:01.140395 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 23 23:23:01.142317 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Nov 23 23:23:01.143870 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Nov 23 23:23:01.149722 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Nov 23 23:23:01.152444 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Nov 23 23:23:01.155151 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Nov 23 23:23:01.163883 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 23 23:23:01.178234 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Nov 23 23:23:01.188637 systemd-tmpfiles[1206]: ACLs are not supported, ignoring. Nov 23 23:23:01.189033 systemd-tmpfiles[1206]: ACLs are not supported, ignoring. Nov 23 23:23:01.189253 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Nov 23 23:23:01.193346 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 23 23:23:01.198225 systemd[1]: Starting systemd-sysusers.service - Create System Users... Nov 23 23:23:01.214640 kernel: loop1: detected capacity change from 0 to 119840 Nov 23 23:23:01.230022 systemd[1]: Finished systemd-sysusers.service - Create System Users. Nov 23 23:23:01.232471 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 23 23:23:01.237611 kernel: loop2: detected capacity change from 0 to 100632 Nov 23 23:23:01.254240 systemd-tmpfiles[1225]: ACLs are not supported, ignoring. Nov 23 23:23:01.254564 systemd-tmpfiles[1225]: ACLs are not supported, ignoring. Nov 23 23:23:01.258038 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 23 23:23:01.261739 kernel: loop3: detected capacity change from 0 to 207008 Nov 23 23:23:01.269512 kernel: loop4: detected capacity change from 0 to 119840 Nov 23 23:23:01.276534 kernel: loop5: detected capacity change from 0 to 100632 Nov 23 23:23:01.280688 (sd-merge)[1229]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Nov 23 23:23:01.281106 (sd-merge)[1229]: Merged extensions into '/usr'. Nov 23 23:23:01.284735 systemd[1]: Reload requested from client PID 1204 ('systemd-sysext') (unit systemd-sysext.service)... Nov 23 23:23:01.284756 systemd[1]: Reloading... Nov 23 23:23:01.331501 zram_generator::config[1251]: No configuration found. Nov 23 23:23:01.400983 ldconfig[1199]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Nov 23 23:23:01.483904 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Nov 23 23:23:01.484353 systemd[1]: Reloading finished in 199 ms. Nov 23 23:23:01.514192 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Nov 23 23:23:01.515772 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Nov 23 23:23:01.530676 systemd[1]: Starting ensure-sysext.service... Nov 23 23:23:01.532777 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 23 23:23:01.541896 systemd[1]: Reload requested from client PID 1289 ('systemctl') (unit ensure-sysext.service)... Nov 23 23:23:01.541913 systemd[1]: Reloading... Nov 23 23:23:01.551667 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Nov 23 23:23:01.551710 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Nov 23 23:23:01.551942 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Nov 23 23:23:01.552157 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Nov 23 23:23:01.552932 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Nov 23 23:23:01.553169 systemd-tmpfiles[1290]: ACLs are not supported, ignoring. Nov 23 23:23:01.553218 systemd-tmpfiles[1290]: ACLs are not supported, ignoring. Nov 23 23:23:01.556194 systemd-tmpfiles[1290]: Detected autofs mount point /boot during canonicalization of boot. Nov 23 23:23:01.556206 systemd-tmpfiles[1290]: Skipping /boot Nov 23 23:23:01.562049 systemd-tmpfiles[1290]: Detected autofs mount point /boot during canonicalization of boot. Nov 23 23:23:01.562064 systemd-tmpfiles[1290]: Skipping /boot Nov 23 23:23:01.576515 zram_generator::config[1316]: No configuration found. Nov 23 23:23:01.711776 systemd[1]: Reloading finished in 169 ms. Nov 23 23:23:01.735988 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Nov 23 23:23:01.742099 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 23 23:23:01.747899 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 23 23:23:01.750465 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Nov 23 23:23:01.761678 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Nov 23 23:23:01.764852 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 23 23:23:01.767852 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 23 23:23:01.770198 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Nov 23 23:23:01.776990 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 23 23:23:01.783777 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 23 23:23:01.786346 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 23 23:23:01.790186 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 23 23:23:01.791660 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 23 23:23:01.791781 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 23 23:23:01.794188 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Nov 23 23:23:01.801522 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Nov 23 23:23:01.803362 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 23 23:23:01.805924 augenrules[1382]: No rules Nov 23 23:23:01.806046 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 23 23:23:01.809151 systemd[1]: audit-rules.service: Deactivated successfully. Nov 23 23:23:01.809340 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 23 23:23:01.810919 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 23 23:23:01.811105 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 23 23:23:01.812694 systemd-udevd[1361]: Using default interface naming scheme 'v255'. Nov 23 23:23:01.812884 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 23 23:23:01.813057 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 23 23:23:01.818448 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Nov 23 23:23:01.823934 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Nov 23 23:23:01.829200 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 23 23:23:01.830351 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 23 23:23:01.832616 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 23 23:23:01.834509 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 23 23:23:01.844658 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 23 23:23:01.849169 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 23 23:23:01.851376 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 23 23:23:01.851420 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 23 23:23:01.854651 systemd[1]: Starting systemd-update-done.service - Update is Completed... Nov 23 23:23:01.855717 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Nov 23 23:23:01.857030 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 23 23:23:01.860755 systemd[1]: Started systemd-userdbd.service - User Database Manager. Nov 23 23:23:01.862329 systemd[1]: Finished ensure-sysext.service. Nov 23 23:23:01.863775 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 23 23:23:01.863933 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 23 23:23:01.865903 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 23 23:23:01.866037 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 23 23:23:01.868898 systemd[1]: Finished systemd-update-done.service - Update is Completed. Nov 23 23:23:01.889676 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 23 23:23:01.894578 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Nov 23 23:23:01.896013 augenrules[1394]: /sbin/augenrules: No change Nov 23 23:23:01.899086 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 23 23:23:01.900558 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 23 23:23:01.903996 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 23 23:23:01.904371 augenrules[1456]: No rules Nov 23 23:23:01.905560 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 23 23:23:01.908808 systemd[1]: audit-rules.service: Deactivated successfully. Nov 23 23:23:01.908997 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 23 23:23:01.921048 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Nov 23 23:23:01.928527 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 23 23:23:01.928584 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 23 23:23:01.940540 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Nov 23 23:23:01.943885 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Nov 23 23:23:01.973687 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Nov 23 23:23:02.000323 systemd-resolved[1357]: Positive Trust Anchors: Nov 23 23:23:02.000338 systemd-resolved[1357]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 23 23:23:02.000369 systemd-resolved[1357]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 23 23:23:02.000837 systemd-networkd[1446]: lo: Link UP Nov 23 23:23:02.000840 systemd-networkd[1446]: lo: Gained carrier Nov 23 23:23:02.001726 systemd-networkd[1446]: Enumeration completed Nov 23 23:23:02.001825 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 23 23:23:02.002240 systemd-networkd[1446]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 23 23:23:02.002244 systemd-networkd[1446]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 23 23:23:02.002784 systemd-networkd[1446]: eth0: Link UP Nov 23 23:23:02.002879 systemd-networkd[1446]: eth0: Gained carrier Nov 23 23:23:02.002892 systemd-networkd[1446]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 23 23:23:02.004322 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Nov 23 23:23:02.007318 systemd-resolved[1357]: Defaulting to hostname 'linux'. Nov 23 23:23:02.007900 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Nov 23 23:23:02.009585 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 23 23:23:02.011167 systemd[1]: Reached target network.target - Network. Nov 23 23:23:02.012116 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 23 23:23:02.016114 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Nov 23 23:23:02.016571 systemd-networkd[1446]: eth0: DHCPv4 address 10.0.0.134/16, gateway 10.0.0.1 acquired from 10.0.0.1 Nov 23 23:23:02.017519 systemd[1]: Reached target sysinit.target - System Initialization. Nov 23 23:23:02.018659 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Nov 23 23:23:02.019984 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Nov 23 23:23:02.020381 systemd-timesyncd[1447]: Network configuration changed, trying to establish connection. Nov 23 23:23:02.021396 systemd-timesyncd[1447]: Contacted time server 10.0.0.1:123 (10.0.0.1). Nov 23 23:23:02.021446 systemd-timesyncd[1447]: Initial clock synchronization to Sun 2025-11-23 23:23:01.895962 UTC. Nov 23 23:23:02.021574 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Nov 23 23:23:02.023060 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Nov 23 23:23:02.023100 systemd[1]: Reached target paths.target - Path Units. Nov 23 23:23:02.024022 systemd[1]: Reached target time-set.target - System Time Set. Nov 23 23:23:02.025214 systemd[1]: Started logrotate.timer - Daily rotation of log files. Nov 23 23:23:02.026591 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Nov 23 23:23:02.027800 systemd[1]: Reached target timers.target - Timer Units. Nov 23 23:23:02.029351 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Nov 23 23:23:02.031633 systemd[1]: Starting docker.socket - Docker Socket for the API... Nov 23 23:23:02.034016 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Nov 23 23:23:02.035429 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Nov 23 23:23:02.036752 systemd[1]: Reached target ssh-access.target - SSH Access Available. Nov 23 23:23:02.047130 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Nov 23 23:23:02.048683 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Nov 23 23:23:02.050741 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Nov 23 23:23:02.052167 systemd[1]: Listening on docker.socket - Docker Socket for the API. Nov 23 23:23:02.053725 systemd[1]: Reached target sockets.target - Socket Units. Nov 23 23:23:02.054714 systemd[1]: Reached target basic.target - Basic System. Nov 23 23:23:02.055700 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Nov 23 23:23:02.055725 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Nov 23 23:23:02.056755 systemd[1]: Starting containerd.service - containerd container runtime... Nov 23 23:23:02.060868 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Nov 23 23:23:02.062674 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Nov 23 23:23:02.067493 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Nov 23 23:23:02.075630 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Nov 23 23:23:02.076787 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Nov 23 23:23:02.077586 jq[1497]: false Nov 23 23:23:02.077949 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Nov 23 23:23:02.083577 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Nov 23 23:23:02.085502 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Nov 23 23:23:02.087665 extend-filesystems[1498]: Found /dev/vda6 Nov 23 23:23:02.090643 extend-filesystems[1498]: Found /dev/vda9 Nov 23 23:23:02.090676 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Nov 23 23:23:02.095858 systemd[1]: Starting systemd-logind.service - User Login Management... Nov 23 23:23:02.097757 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Nov 23 23:23:02.098140 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Nov 23 23:23:02.103865 extend-filesystems[1498]: Checking size of /dev/vda9 Nov 23 23:23:02.100620 systemd[1]: Starting update-engine.service - Update Engine... Nov 23 23:23:02.102699 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Nov 23 23:23:02.108046 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Nov 23 23:23:02.110462 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Nov 23 23:23:02.112520 jq[1516]: true Nov 23 23:23:02.110688 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Nov 23 23:23:02.110916 systemd[1]: motdgen.service: Deactivated successfully. Nov 23 23:23:02.111125 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Nov 23 23:23:02.115989 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Nov 23 23:23:02.116180 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Nov 23 23:23:02.116421 extend-filesystems[1498]: Resized partition /dev/vda9 Nov 23 23:23:02.126525 extend-filesystems[1524]: resize2fs 1.47.3 (8-Jul-2025) Nov 23 23:23:02.136050 jq[1526]: true Nov 23 23:23:02.142557 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Nov 23 23:23:02.145216 tar[1522]: linux-arm64/LICENSE Nov 23 23:23:02.145216 tar[1522]: linux-arm64/helm Nov 23 23:23:02.154175 (ntainerd)[1527]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Nov 23 23:23:02.158152 update_engine[1514]: I20251123 23:23:02.155968 1514 main.cc:92] Flatcar Update Engine starting Nov 23 23:23:02.167106 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Nov 23 23:23:02.166588 systemd[1]: Started dbus.service - D-Bus System Message Bus. Nov 23 23:23:02.166354 dbus-daemon[1495]: [system] SELinux support is enabled Nov 23 23:23:02.180899 update_engine[1514]: I20251123 23:23:02.172632 1514 update_check_scheduler.cc:74] Next update check in 8m32s Nov 23 23:23:02.174847 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Nov 23 23:23:02.174870 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Nov 23 23:23:02.179653 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 23 23:23:02.181082 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Nov 23 23:23:02.181109 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Nov 23 23:23:02.182621 systemd[1]: Started update-engine.service - Update Engine. Nov 23 23:23:02.182974 extend-filesystems[1524]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Nov 23 23:23:02.182974 extend-filesystems[1524]: old_desc_blocks = 1, new_desc_blocks = 1 Nov 23 23:23:02.182974 extend-filesystems[1524]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Nov 23 23:23:02.191965 extend-filesystems[1498]: Resized filesystem in /dev/vda9 Nov 23 23:23:02.183943 systemd[1]: extend-filesystems.service: Deactivated successfully. Nov 23 23:23:02.185529 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Nov 23 23:23:02.193596 systemd[1]: Started locksmithd.service - Cluster reboot manager. Nov 23 23:23:02.207797 bash[1558]: Updated "/home/core/.ssh/authorized_keys" Nov 23 23:23:02.208850 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Nov 23 23:23:02.213874 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Nov 23 23:23:02.217363 systemd-logind[1513]: Watching system buttons on /dev/input/event0 (Power Button) Nov 23 23:23:02.217820 systemd-logind[1513]: New seat seat0. Nov 23 23:23:02.219734 systemd[1]: Started systemd-logind.service - User Login Management. Nov 23 23:23:02.245552 locksmithd[1557]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Nov 23 23:23:02.258234 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 23 23:23:02.331148 containerd[1527]: time="2025-11-23T23:23:02Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Nov 23 23:23:02.332186 containerd[1527]: time="2025-11-23T23:23:02.332134840Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Nov 23 23:23:02.341872 containerd[1527]: time="2025-11-23T23:23:02.341816440Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="9.6µs" Nov 23 23:23:02.341872 containerd[1527]: time="2025-11-23T23:23:02.341849560Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Nov 23 23:23:02.341872 containerd[1527]: time="2025-11-23T23:23:02.341870440Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Nov 23 23:23:02.342024 containerd[1527]: time="2025-11-23T23:23:02.342004640Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Nov 23 23:23:02.342046 containerd[1527]: time="2025-11-23T23:23:02.342030600Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Nov 23 23:23:02.342062 containerd[1527]: time="2025-11-23T23:23:02.342052680Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342127 containerd[1527]: time="2025-11-23T23:23:02.342110080Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342154 containerd[1527]: time="2025-11-23T23:23:02.342126000Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342430 containerd[1527]: time="2025-11-23T23:23:02.342408960Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342450 containerd[1527]: time="2025-11-23T23:23:02.342430520Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342450 containerd[1527]: time="2025-11-23T23:23:02.342441000Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342496 containerd[1527]: time="2025-11-23T23:23:02.342448720Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342567 containerd[1527]: time="2025-11-23T23:23:02.342549120Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342744 containerd[1527]: time="2025-11-23T23:23:02.342726160Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342775 containerd[1527]: time="2025-11-23T23:23:02.342761240Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Nov 23 23:23:02.342798 containerd[1527]: time="2025-11-23T23:23:02.342774640Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Nov 23 23:23:02.342828 containerd[1527]: time="2025-11-23T23:23:02.342816920Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Nov 23 23:23:02.343078 containerd[1527]: time="2025-11-23T23:23:02.343054120Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Nov 23 23:23:02.343145 containerd[1527]: time="2025-11-23T23:23:02.343130680Z" level=info msg="metadata content store policy set" policy=shared Nov 23 23:23:02.346642 containerd[1527]: time="2025-11-23T23:23:02.346615720Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Nov 23 23:23:02.346709 containerd[1527]: time="2025-11-23T23:23:02.346663880Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Nov 23 23:23:02.346709 containerd[1527]: time="2025-11-23T23:23:02.346685000Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Nov 23 23:23:02.346709 containerd[1527]: time="2025-11-23T23:23:02.346696640Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Nov 23 23:23:02.346709 containerd[1527]: time="2025-11-23T23:23:02.346708240Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Nov 23 23:23:02.346771 containerd[1527]: time="2025-11-23T23:23:02.346721560Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Nov 23 23:23:02.346771 containerd[1527]: time="2025-11-23T23:23:02.346733280Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Nov 23 23:23:02.346771 containerd[1527]: time="2025-11-23T23:23:02.346743960Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Nov 23 23:23:02.346771 containerd[1527]: time="2025-11-23T23:23:02.346755800Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Nov 23 23:23:02.346771 containerd[1527]: time="2025-11-23T23:23:02.346767000Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Nov 23 23:23:02.346848 containerd[1527]: time="2025-11-23T23:23:02.346775760Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Nov 23 23:23:02.346848 containerd[1527]: time="2025-11-23T23:23:02.346787320Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Nov 23 23:23:02.346952 containerd[1527]: time="2025-11-23T23:23:02.346917160Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Nov 23 23:23:02.346986 containerd[1527]: time="2025-11-23T23:23:02.346958280Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Nov 23 23:23:02.346986 containerd[1527]: time="2025-11-23T23:23:02.346972320Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Nov 23 23:23:02.346986 containerd[1527]: time="2025-11-23T23:23:02.346983840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Nov 23 23:23:02.347032 containerd[1527]: time="2025-11-23T23:23:02.346995400Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Nov 23 23:23:02.347032 containerd[1527]: time="2025-11-23T23:23:02.347005200Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Nov 23 23:23:02.347032 containerd[1527]: time="2025-11-23T23:23:02.347015560Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Nov 23 23:23:02.347032 containerd[1527]: time="2025-11-23T23:23:02.347024720Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Nov 23 23:23:02.347111 containerd[1527]: time="2025-11-23T23:23:02.347035760Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Nov 23 23:23:02.347111 containerd[1527]: time="2025-11-23T23:23:02.347048400Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Nov 23 23:23:02.347111 containerd[1527]: time="2025-11-23T23:23:02.347058160Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Nov 23 23:23:02.347252 containerd[1527]: time="2025-11-23T23:23:02.347233680Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Nov 23 23:23:02.347288 containerd[1527]: time="2025-11-23T23:23:02.347253400Z" level=info msg="Start snapshots syncer" Nov 23 23:23:02.347288 containerd[1527]: time="2025-11-23T23:23:02.347277880Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Nov 23 23:23:02.347574 containerd[1527]: time="2025-11-23T23:23:02.347534480Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Nov 23 23:23:02.347679 containerd[1527]: time="2025-11-23T23:23:02.347589920Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Nov 23 23:23:02.347679 containerd[1527]: time="2025-11-23T23:23:02.347633840Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Nov 23 23:23:02.347747 containerd[1527]: time="2025-11-23T23:23:02.347728240Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Nov 23 23:23:02.347771 containerd[1527]: time="2025-11-23T23:23:02.347754160Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Nov 23 23:23:02.347771 containerd[1527]: time="2025-11-23T23:23:02.347765960Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Nov 23 23:23:02.347912 containerd[1527]: time="2025-11-23T23:23:02.347777680Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Nov 23 23:23:02.347912 containerd[1527]: time="2025-11-23T23:23:02.347789760Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Nov 23 23:23:02.347912 containerd[1527]: time="2025-11-23T23:23:02.347818920Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Nov 23 23:23:02.347912 containerd[1527]: time="2025-11-23T23:23:02.347838440Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Nov 23 23:23:02.347912 containerd[1527]: time="2025-11-23T23:23:02.347864280Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Nov 23 23:23:02.347912 containerd[1527]: time="2025-11-23T23:23:02.347876120Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Nov 23 23:23:02.347912 containerd[1527]: time="2025-11-23T23:23:02.347886240Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Nov 23 23:23:02.348034 containerd[1527]: time="2025-11-23T23:23:02.347924160Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Nov 23 23:23:02.348034 containerd[1527]: time="2025-11-23T23:23:02.347939360Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Nov 23 23:23:02.348034 containerd[1527]: time="2025-11-23T23:23:02.347947920Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Nov 23 23:23:02.348034 containerd[1527]: time="2025-11-23T23:23:02.347957200Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Nov 23 23:23:02.348034 containerd[1527]: time="2025-11-23T23:23:02.347964600Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Nov 23 23:23:02.348034 containerd[1527]: time="2025-11-23T23:23:02.347972960Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Nov 23 23:23:02.348034 containerd[1527]: time="2025-11-23T23:23:02.347982560Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Nov 23 23:23:02.348157 containerd[1527]: time="2025-11-23T23:23:02.348056080Z" level=info msg="runtime interface created" Nov 23 23:23:02.348157 containerd[1527]: time="2025-11-23T23:23:02.348061520Z" level=info msg="created NRI interface" Nov 23 23:23:02.348157 containerd[1527]: time="2025-11-23T23:23:02.348076960Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Nov 23 23:23:02.348157 containerd[1527]: time="2025-11-23T23:23:02.348089360Z" level=info msg="Connect containerd service" Nov 23 23:23:02.348157 containerd[1527]: time="2025-11-23T23:23:02.348110080Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Nov 23 23:23:02.348946 containerd[1527]: time="2025-11-23T23:23:02.348892960Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 23 23:23:02.417326 containerd[1527]: time="2025-11-23T23:23:02.417240120Z" level=info msg="Start subscribing containerd event" Nov 23 23:23:02.417446 containerd[1527]: time="2025-11-23T23:23:02.417429960Z" level=info msg="Start recovering state" Nov 23 23:23:02.417736 containerd[1527]: time="2025-11-23T23:23:02.417633080Z" level=info msg="Start event monitor" Nov 23 23:23:02.417848 containerd[1527]: time="2025-11-23T23:23:02.417831800Z" level=info msg="Start cni network conf syncer for default" Nov 23 23:23:02.417870 containerd[1527]: time="2025-11-23T23:23:02.417848240Z" level=info msg="Start streaming server" Nov 23 23:23:02.417870 containerd[1527]: time="2025-11-23T23:23:02.417858880Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Nov 23 23:23:02.417870 containerd[1527]: time="2025-11-23T23:23:02.417866200Z" level=info msg="runtime interface starting up..." Nov 23 23:23:02.417917 containerd[1527]: time="2025-11-23T23:23:02.417871800Z" level=info msg="starting plugins..." Nov 23 23:23:02.418356 containerd[1527]: time="2025-11-23T23:23:02.418312760Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Nov 23 23:23:02.418401 containerd[1527]: time="2025-11-23T23:23:02.418388320Z" level=info msg=serving... address=/run/containerd/containerd.sock Nov 23 23:23:02.420507 containerd[1527]: time="2025-11-23T23:23:02.420392720Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Nov 23 23:23:02.420737 containerd[1527]: time="2025-11-23T23:23:02.420719120Z" level=info msg="containerd successfully booted in 0.089919s" Nov 23 23:23:02.420821 systemd[1]: Started containerd.service - containerd container runtime. Nov 23 23:23:02.494593 tar[1522]: linux-arm64/README.md Nov 23 23:23:02.510725 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Nov 23 23:23:02.771638 sshd_keygen[1525]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Nov 23 23:23:02.790529 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Nov 23 23:23:02.793200 systemd[1]: Starting issuegen.service - Generate /run/issue... Nov 23 23:23:02.814061 systemd[1]: issuegen.service: Deactivated successfully. Nov 23 23:23:02.814554 systemd[1]: Finished issuegen.service - Generate /run/issue. Nov 23 23:23:02.817153 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Nov 23 23:23:02.847583 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Nov 23 23:23:02.850365 systemd[1]: Started getty@tty1.service - Getty on tty1. Nov 23 23:23:02.852567 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Nov 23 23:23:02.853909 systemd[1]: Reached target getty.target - Login Prompts. Nov 23 23:23:03.395652 systemd-networkd[1446]: eth0: Gained IPv6LL Nov 23 23:23:03.397884 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Nov 23 23:23:03.399614 systemd[1]: Reached target network-online.target - Network is Online. Nov 23 23:23:03.401876 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Nov 23 23:23:03.404195 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 23 23:23:03.412961 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Nov 23 23:23:03.426113 systemd[1]: coreos-metadata.service: Deactivated successfully. Nov 23 23:23:03.426325 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Nov 23 23:23:03.428002 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Nov 23 23:23:03.430365 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Nov 23 23:23:03.948344 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 23 23:23:03.949959 systemd[1]: Reached target multi-user.target - Multi-User System. Nov 23 23:23:03.952061 (kubelet)[1634]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 23 23:23:03.956783 systemd[1]: Startup finished in 2.110s (kernel) + 4.841s (initrd) + 3.560s (userspace) = 10.512s. Nov 23 23:23:04.285435 kubelet[1634]: E1123 23:23:04.285306 1634 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 23 23:23:04.287717 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 23 23:23:04.287846 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 23 23:23:04.288168 systemd[1]: kubelet.service: Consumed 744ms CPU time, 256.8M memory peak. Nov 23 23:23:08.500046 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Nov 23 23:23:08.501009 systemd[1]: Started sshd@0-10.0.0.134:22-10.0.0.1:54788.service - OpenSSH per-connection server daemon (10.0.0.1:54788). Nov 23 23:23:08.582928 sshd[1647]: Accepted publickey for core from 10.0.0.1 port 54788 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:23:08.584903 sshd-session[1647]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:23:08.591691 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Nov 23 23:23:08.592640 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Nov 23 23:23:08.598180 systemd-logind[1513]: New session 1 of user core. Nov 23 23:23:08.611524 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Nov 23 23:23:08.613861 systemd[1]: Starting user@500.service - User Manager for UID 500... Nov 23 23:23:08.633363 (systemd)[1652]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Nov 23 23:23:08.635388 systemd-logind[1513]: New session c1 of user core. Nov 23 23:23:08.743317 systemd[1652]: Queued start job for default target default.target. Nov 23 23:23:08.765417 systemd[1652]: Created slice app.slice - User Application Slice. Nov 23 23:23:08.765445 systemd[1652]: Reached target paths.target - Paths. Nov 23 23:23:08.765511 systemd[1652]: Reached target timers.target - Timers. Nov 23 23:23:08.766719 systemd[1652]: Starting dbus.socket - D-Bus User Message Bus Socket... Nov 23 23:23:08.781779 systemd[1652]: Listening on dbus.socket - D-Bus User Message Bus Socket. Nov 23 23:23:08.781910 systemd[1652]: Reached target sockets.target - Sockets. Nov 23 23:23:08.781960 systemd[1652]: Reached target basic.target - Basic System. Nov 23 23:23:08.781989 systemd[1652]: Reached target default.target - Main User Target. Nov 23 23:23:08.782013 systemd[1652]: Startup finished in 141ms. Nov 23 23:23:08.782282 systemd[1]: Started user@500.service - User Manager for UID 500. Nov 23 23:23:08.785773 systemd[1]: Started session-1.scope - Session 1 of User core. Nov 23 23:23:08.847311 systemd[1]: Started sshd@1-10.0.0.134:22-10.0.0.1:54806.service - OpenSSH per-connection server daemon (10.0.0.1:54806). Nov 23 23:23:08.920878 sshd[1663]: Accepted publickey for core from 10.0.0.1 port 54806 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:23:08.922138 sshd-session[1663]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:23:08.926537 systemd-logind[1513]: New session 2 of user core. Nov 23 23:23:08.945645 systemd[1]: Started session-2.scope - Session 2 of User core. Nov 23 23:23:08.996011 sshd[1666]: Connection closed by 10.0.0.1 port 54806 Nov 23 23:23:08.996345 sshd-session[1663]: pam_unix(sshd:session): session closed for user core Nov 23 23:23:09.017374 systemd[1]: sshd@1-10.0.0.134:22-10.0.0.1:54806.service: Deactivated successfully. Nov 23 23:23:09.018861 systemd[1]: session-2.scope: Deactivated successfully. Nov 23 23:23:09.019631 systemd-logind[1513]: Session 2 logged out. Waiting for processes to exit. Nov 23 23:23:09.021409 systemd[1]: Started sshd@2-10.0.0.134:22-10.0.0.1:54822.service - OpenSSH per-connection server daemon (10.0.0.1:54822). Nov 23 23:23:09.022178 systemd-logind[1513]: Removed session 2. Nov 23 23:23:09.084918 sshd[1672]: Accepted publickey for core from 10.0.0.1 port 54822 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:23:09.085955 sshd-session[1672]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:23:09.089798 systemd-logind[1513]: New session 3 of user core. Nov 23 23:23:09.098623 systemd[1]: Started session-3.scope - Session 3 of User core. Nov 23 23:23:09.148026 sshd[1675]: Connection closed by 10.0.0.1 port 54822 Nov 23 23:23:09.147885 sshd-session[1672]: pam_unix(sshd:session): session closed for user core Nov 23 23:23:09.157465 systemd[1]: sshd@2-10.0.0.134:22-10.0.0.1:54822.service: Deactivated successfully. Nov 23 23:23:09.158855 systemd[1]: session-3.scope: Deactivated successfully. Nov 23 23:23:09.160171 systemd-logind[1513]: Session 3 logged out. Waiting for processes to exit. Nov 23 23:23:09.161231 systemd[1]: Started sshd@3-10.0.0.134:22-10.0.0.1:54830.service - OpenSSH per-connection server daemon (10.0.0.1:54830). Nov 23 23:23:09.161954 systemd-logind[1513]: Removed session 3. Nov 23 23:23:09.209663 sshd[1681]: Accepted publickey for core from 10.0.0.1 port 54830 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:23:09.210776 sshd-session[1681]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:23:09.214574 systemd-logind[1513]: New session 4 of user core. Nov 23 23:23:09.232654 systemd[1]: Started session-4.scope - Session 4 of User core. Nov 23 23:23:09.283079 sshd[1684]: Connection closed by 10.0.0.1 port 54830 Nov 23 23:23:09.283339 sshd-session[1681]: pam_unix(sshd:session): session closed for user core Nov 23 23:23:09.293302 systemd[1]: sshd@3-10.0.0.134:22-10.0.0.1:54830.service: Deactivated successfully. Nov 23 23:23:09.294685 systemd[1]: session-4.scope: Deactivated successfully. Nov 23 23:23:09.295825 systemd-logind[1513]: Session 4 logged out. Waiting for processes to exit. Nov 23 23:23:09.297273 systemd[1]: Started sshd@4-10.0.0.134:22-10.0.0.1:54850.service - OpenSSH per-connection server daemon (10.0.0.1:54850). Nov 23 23:23:09.298350 systemd-logind[1513]: Removed session 4. Nov 23 23:23:09.364065 sshd[1690]: Accepted publickey for core from 10.0.0.1 port 54850 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:23:09.365282 sshd-session[1690]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:23:09.369921 systemd-logind[1513]: New session 5 of user core. Nov 23 23:23:09.385653 systemd[1]: Started session-5.scope - Session 5 of User core. Nov 23 23:23:09.441114 sudo[1694]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Nov 23 23:23:09.441362 sudo[1694]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 23 23:23:09.456509 sudo[1694]: pam_unix(sudo:session): session closed for user root Nov 23 23:23:09.458286 sshd[1693]: Connection closed by 10.0.0.1 port 54850 Nov 23 23:23:09.458126 sshd-session[1690]: pam_unix(sshd:session): session closed for user core Nov 23 23:23:09.467352 systemd[1]: sshd@4-10.0.0.134:22-10.0.0.1:54850.service: Deactivated successfully. Nov 23 23:23:09.469688 systemd[1]: session-5.scope: Deactivated successfully. Nov 23 23:23:09.472642 systemd-logind[1513]: Session 5 logged out. Waiting for processes to exit. Nov 23 23:23:09.475367 systemd-logind[1513]: Removed session 5. Nov 23 23:23:09.476982 systemd[1]: Started sshd@5-10.0.0.134:22-10.0.0.1:60718.service - OpenSSH per-connection server daemon (10.0.0.1:60718). Nov 23 23:23:09.526745 sshd[1700]: Accepted publickey for core from 10.0.0.1 port 60718 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:23:09.527975 sshd-session[1700]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:23:09.532502 systemd-logind[1513]: New session 6 of user core. Nov 23 23:23:09.548657 systemd[1]: Started session-6.scope - Session 6 of User core. Nov 23 23:23:09.600691 sudo[1705]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Nov 23 23:23:09.601497 sudo[1705]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 23 23:23:09.676583 sudo[1705]: pam_unix(sudo:session): session closed for user root Nov 23 23:23:09.681371 sudo[1704]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Nov 23 23:23:09.681651 sudo[1704]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 23 23:23:09.691469 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 23 23:23:09.732683 augenrules[1727]: No rules Nov 23 23:23:09.733349 systemd[1]: audit-rules.service: Deactivated successfully. Nov 23 23:23:09.733630 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 23 23:23:09.734763 sudo[1704]: pam_unix(sudo:session): session closed for user root Nov 23 23:23:09.736070 sshd[1703]: Connection closed by 10.0.0.1 port 60718 Nov 23 23:23:09.736365 sshd-session[1700]: pam_unix(sshd:session): session closed for user core Nov 23 23:23:09.749488 systemd[1]: sshd@5-10.0.0.134:22-10.0.0.1:60718.service: Deactivated successfully. Nov 23 23:23:09.752745 systemd[1]: session-6.scope: Deactivated successfully. Nov 23 23:23:09.753408 systemd-logind[1513]: Session 6 logged out. Waiting for processes to exit. Nov 23 23:23:09.755458 systemd[1]: Started sshd@6-10.0.0.134:22-10.0.0.1:60732.service - OpenSSH per-connection server daemon (10.0.0.1:60732). Nov 23 23:23:09.756155 systemd-logind[1513]: Removed session 6. Nov 23 23:23:09.819698 sshd[1736]: Accepted publickey for core from 10.0.0.1 port 60732 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:23:09.820532 sshd-session[1736]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:23:09.824556 systemd-logind[1513]: New session 7 of user core. Nov 23 23:23:09.835644 systemd[1]: Started session-7.scope - Session 7 of User core. Nov 23 23:23:09.886120 sudo[1741]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Nov 23 23:23:09.886383 sudo[1741]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 23 23:23:10.166717 systemd[1]: Starting docker.service - Docker Application Container Engine... Nov 23 23:23:10.184758 (dockerd)[1762]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Nov 23 23:23:10.383212 dockerd[1762]: time="2025-11-23T23:23:10.382849366Z" level=info msg="Starting up" Nov 23 23:23:10.383910 dockerd[1762]: time="2025-11-23T23:23:10.383876208Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Nov 23 23:23:10.394226 dockerd[1762]: time="2025-11-23T23:23:10.394195002Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Nov 23 23:23:10.428006 dockerd[1762]: time="2025-11-23T23:23:10.427906724Z" level=info msg="Loading containers: start." Nov 23 23:23:10.436509 kernel: Initializing XFRM netlink socket Nov 23 23:23:10.630017 systemd-networkd[1446]: docker0: Link UP Nov 23 23:23:10.635265 dockerd[1762]: time="2025-11-23T23:23:10.635205087Z" level=info msg="Loading containers: done." Nov 23 23:23:10.649590 dockerd[1762]: time="2025-11-23T23:23:10.649535166Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Nov 23 23:23:10.649734 dockerd[1762]: time="2025-11-23T23:23:10.649613918Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Nov 23 23:23:10.649734 dockerd[1762]: time="2025-11-23T23:23:10.649692711Z" level=info msg="Initializing buildkit" Nov 23 23:23:10.672928 dockerd[1762]: time="2025-11-23T23:23:10.672879118Z" level=info msg="Completed buildkit initialization" Nov 23 23:23:10.677470 dockerd[1762]: time="2025-11-23T23:23:10.677427915Z" level=info msg="Daemon has completed initialization" Nov 23 23:23:10.677559 dockerd[1762]: time="2025-11-23T23:23:10.677491680Z" level=info msg="API listen on /run/docker.sock" Nov 23 23:23:10.677736 systemd[1]: Started docker.service - Docker Application Container Engine. Nov 23 23:23:11.161710 containerd[1527]: time="2025-11-23T23:23:11.161672398Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.10\"" Nov 23 23:23:11.686518 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount602846265.mount: Deactivated successfully. Nov 23 23:23:12.668965 containerd[1527]: time="2025-11-23T23:23:12.668889719Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:12.669547 containerd[1527]: time="2025-11-23T23:23:12.669466430Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.10: active requests=0, bytes read=26431961" Nov 23 23:23:12.670526 containerd[1527]: time="2025-11-23T23:23:12.670475175Z" level=info msg="ImageCreate event name:\"sha256:03aec5fd5841efdd990b8fe285e036fc1386e2f8851378ce2c9dfd1b331897ea\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:12.673500 containerd[1527]: time="2025-11-23T23:23:12.673161022Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:af4ee57c047e31a7f58422b94a9ec4c62221d3deebb16755bdeff720df796189\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:12.674339 containerd[1527]: time="2025-11-23T23:23:12.674183582Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.10\" with image id \"sha256:03aec5fd5841efdd990b8fe285e036fc1386e2f8851378ce2c9dfd1b331897ea\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:af4ee57c047e31a7f58422b94a9ec4c62221d3deebb16755bdeff720df796189\", size \"26428558\" in 1.512470996s" Nov 23 23:23:12.674339 containerd[1527]: time="2025-11-23T23:23:12.674218218Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.10\" returns image reference \"sha256:03aec5fd5841efdd990b8fe285e036fc1386e2f8851378ce2c9dfd1b331897ea\"" Nov 23 23:23:12.674908 containerd[1527]: time="2025-11-23T23:23:12.674878652Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.10\"" Nov 23 23:23:13.723071 containerd[1527]: time="2025-11-23T23:23:13.723022297Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:13.724047 containerd[1527]: time="2025-11-23T23:23:13.724008654Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.10: active requests=0, bytes read=22618957" Nov 23 23:23:13.725130 containerd[1527]: time="2025-11-23T23:23:13.724739549Z" level=info msg="ImageCreate event name:\"sha256:66490a6490dde2df4a78eba21320da67070ad88461899536880edb5301ec2ba3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:13.727933 containerd[1527]: time="2025-11-23T23:23:13.727905006Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:efbd9d1dfcd2940e1c73a1476c880c3c2cdf04cc60722d329b21cd48745c8660\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:13.729498 containerd[1527]: time="2025-11-23T23:23:13.729428421Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.10\" with image id \"sha256:66490a6490dde2df4a78eba21320da67070ad88461899536880edb5301ec2ba3\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:efbd9d1dfcd2940e1c73a1476c880c3c2cdf04cc60722d329b21cd48745c8660\", size \"24203439\" in 1.054520504s" Nov 23 23:23:13.729498 containerd[1527]: time="2025-11-23T23:23:13.729467858Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.10\" returns image reference \"sha256:66490a6490dde2df4a78eba21320da67070ad88461899536880edb5301ec2ba3\"" Nov 23 23:23:13.729902 containerd[1527]: time="2025-11-23T23:23:13.729875928Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.10\"" Nov 23 23:23:14.538209 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Nov 23 23:23:14.540910 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 23 23:23:14.709181 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 23 23:23:14.725913 (kubelet)[2053]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 23 23:23:14.863327 containerd[1527]: time="2025-11-23T23:23:14.863211518Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:14.864344 containerd[1527]: time="2025-11-23T23:23:14.864085993Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.10: active requests=0, bytes read=17618438" Nov 23 23:23:14.869784 containerd[1527]: time="2025-11-23T23:23:14.869743792Z" level=info msg="ImageCreate event name:\"sha256:fcf368a1abd0b48cff2fd3cca12fcc008aaf52eeab885656f11e7773c6a188a3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:14.875376 containerd[1527]: time="2025-11-23T23:23:14.875327778Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:9c58e1adcad5af66d1d9ca5cf9a4c266e4054b8f19f91a8fff1993549e657b10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:14.877320 containerd[1527]: time="2025-11-23T23:23:14.877276723Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.10\" with image id \"sha256:fcf368a1abd0b48cff2fd3cca12fcc008aaf52eeab885656f11e7773c6a188a3\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:9c58e1adcad5af66d1d9ca5cf9a4c266e4054b8f19f91a8fff1993549e657b10\", size \"19202938\" in 1.147311361s" Nov 23 23:23:14.877362 containerd[1527]: time="2025-11-23T23:23:14.877322996Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.10\" returns image reference \"sha256:fcf368a1abd0b48cff2fd3cca12fcc008aaf52eeab885656f11e7773c6a188a3\"" Nov 23 23:23:14.877948 containerd[1527]: time="2025-11-23T23:23:14.877913538Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.10\"" Nov 23 23:23:14.881773 kubelet[2053]: E1123 23:23:14.881727 2053 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 23 23:23:14.884682 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 23 23:23:14.884813 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 23 23:23:14.886566 systemd[1]: kubelet.service: Consumed 263ms CPU time, 107.6M memory peak. Nov 23 23:23:15.853808 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3189627699.mount: Deactivated successfully. Nov 23 23:23:16.191868 containerd[1527]: time="2025-11-23T23:23:16.191737462Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:16.192510 containerd[1527]: time="2025-11-23T23:23:16.192447575Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.10: active requests=0, bytes read=27561801" Nov 23 23:23:16.193425 containerd[1527]: time="2025-11-23T23:23:16.193338668Z" level=info msg="ImageCreate event name:\"sha256:8b57c1f8bd2ddfa793889457b41e87132f192046e262b32ab0514f32d28be47d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:16.196084 containerd[1527]: time="2025-11-23T23:23:16.195879911Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e3dda1c7b384f9eb5b2fa1c27493b23b80e6204b9fa2ee8791b2de078f468cbf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:16.196456 containerd[1527]: time="2025-11-23T23:23:16.196435612Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.10\" with image id \"sha256:8b57c1f8bd2ddfa793889457b41e87132f192046e262b32ab0514f32d28be47d\", repo tag \"registry.k8s.io/kube-proxy:v1.32.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:e3dda1c7b384f9eb5b2fa1c27493b23b80e6204b9fa2ee8791b2de078f468cbf\", size \"27560818\" in 1.318459972s" Nov 23 23:23:16.196507 containerd[1527]: time="2025-11-23T23:23:16.196463415Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.10\" returns image reference \"sha256:8b57c1f8bd2ddfa793889457b41e87132f192046e262b32ab0514f32d28be47d\"" Nov 23 23:23:16.196932 containerd[1527]: time="2025-11-23T23:23:16.196909460Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Nov 23 23:23:16.628552 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2868196043.mount: Deactivated successfully. Nov 23 23:23:17.324247 containerd[1527]: time="2025-11-23T23:23:17.324191653Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:17.325791 containerd[1527]: time="2025-11-23T23:23:17.325755784Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951624" Nov 23 23:23:17.327502 containerd[1527]: time="2025-11-23T23:23:17.327416322Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:17.332500 containerd[1527]: time="2025-11-23T23:23:17.332037409Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:17.334076 containerd[1527]: time="2025-11-23T23:23:17.334030102Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.137089207s" Nov 23 23:23:17.334076 containerd[1527]: time="2025-11-23T23:23:17.334071243Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Nov 23 23:23:17.335285 containerd[1527]: time="2025-11-23T23:23:17.335261719Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Nov 23 23:23:17.753326 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3242364267.mount: Deactivated successfully. Nov 23 23:23:17.758733 containerd[1527]: time="2025-11-23T23:23:17.758687987Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 23 23:23:17.759169 containerd[1527]: time="2025-11-23T23:23:17.759145519Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" Nov 23 23:23:17.760141 containerd[1527]: time="2025-11-23T23:23:17.760113096Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 23 23:23:17.762123 containerd[1527]: time="2025-11-23T23:23:17.762069317Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 23 23:23:17.762951 containerd[1527]: time="2025-11-23T23:23:17.762603942Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 427.251404ms" Nov 23 23:23:17.762951 containerd[1527]: time="2025-11-23T23:23:17.762633311Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Nov 23 23:23:17.763129 containerd[1527]: time="2025-11-23T23:23:17.763083341Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Nov 23 23:23:18.275821 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2497248070.mount: Deactivated successfully. Nov 23 23:23:19.892948 containerd[1527]: time="2025-11-23T23:23:19.892885955Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:19.893855 containerd[1527]: time="2025-11-23T23:23:19.893819584Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=67943167" Nov 23 23:23:19.894507 containerd[1527]: time="2025-11-23T23:23:19.894473293Z" level=info msg="ImageCreate event name:\"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:19.898258 containerd[1527]: time="2025-11-23T23:23:19.898217553Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:19.899193 containerd[1527]: time="2025-11-23T23:23:19.898838323Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"67941650\" in 2.135705856s" Nov 23 23:23:19.899193 containerd[1527]: time="2025-11-23T23:23:19.898877929Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\"" Nov 23 23:23:24.014314 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 23 23:23:24.014459 systemd[1]: kubelet.service: Consumed 263ms CPU time, 107.6M memory peak. Nov 23 23:23:24.016257 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 23 23:23:24.036123 systemd[1]: Reload requested from client PID 2209 ('systemctl') (unit session-7.scope)... Nov 23 23:23:24.036137 systemd[1]: Reloading... Nov 23 23:23:24.104515 zram_generator::config[2252]: No configuration found. Nov 23 23:23:24.257304 systemd[1]: Reloading finished in 220 ms. Nov 23 23:23:24.317958 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Nov 23 23:23:24.318032 systemd[1]: kubelet.service: Failed with result 'signal'. Nov 23 23:23:24.318249 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 23 23:23:24.318300 systemd[1]: kubelet.service: Consumed 90ms CPU time, 95.2M memory peak. Nov 23 23:23:24.320753 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 23 23:23:24.443339 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 23 23:23:24.448603 (kubelet)[2297]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 23 23:23:24.483498 kubelet[2297]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 23:23:24.483498 kubelet[2297]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 23 23:23:24.483498 kubelet[2297]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 23:23:24.483840 kubelet[2297]: I1123 23:23:24.483553 2297 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 23 23:23:25.558299 kubelet[2297]: I1123 23:23:25.558247 2297 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 23 23:23:25.558299 kubelet[2297]: I1123 23:23:25.558284 2297 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 23 23:23:25.558666 kubelet[2297]: I1123 23:23:25.558585 2297 server.go:954] "Client rotation is on, will bootstrap in background" Nov 23 23:23:25.581964 kubelet[2297]: E1123 23:23:25.581917 2297 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.134:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.134:6443: connect: connection refused" logger="UnhandledError" Nov 23 23:23:25.585700 kubelet[2297]: I1123 23:23:25.585660 2297 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 23 23:23:25.591731 kubelet[2297]: I1123 23:23:25.591626 2297 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 23 23:23:25.594849 kubelet[2297]: I1123 23:23:25.594800 2297 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 23 23:23:25.595555 kubelet[2297]: I1123 23:23:25.595506 2297 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 23 23:23:25.595743 kubelet[2297]: I1123 23:23:25.595558 2297 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 23 23:23:25.595847 kubelet[2297]: I1123 23:23:25.595825 2297 topology_manager.go:138] "Creating topology manager with none policy" Nov 23 23:23:25.595847 kubelet[2297]: I1123 23:23:25.595836 2297 container_manager_linux.go:304] "Creating device plugin manager" Nov 23 23:23:25.596048 kubelet[2297]: I1123 23:23:25.596033 2297 state_mem.go:36] "Initialized new in-memory state store" Nov 23 23:23:25.598802 kubelet[2297]: I1123 23:23:25.598757 2297 kubelet.go:446] "Attempting to sync node with API server" Nov 23 23:23:25.598901 kubelet[2297]: I1123 23:23:25.598887 2297 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 23 23:23:25.598926 kubelet[2297]: I1123 23:23:25.598919 2297 kubelet.go:352] "Adding apiserver pod source" Nov 23 23:23:25.599053 kubelet[2297]: I1123 23:23:25.599041 2297 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 23 23:23:25.602611 kubelet[2297]: I1123 23:23:25.602587 2297 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Nov 23 23:23:25.603077 kubelet[2297]: W1123 23:23:25.602952 2297 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.134:6443: connect: connection refused Nov 23 23:23:25.603077 kubelet[2297]: E1123 23:23:25.603015 2297 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.134:6443: connect: connection refused" logger="UnhandledError" Nov 23 23:23:25.603077 kubelet[2297]: W1123 23:23:25.602985 2297 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.134:6443: connect: connection refused Nov 23 23:23:25.603077 kubelet[2297]: E1123 23:23:25.603049 2297 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.134:6443: connect: connection refused" logger="UnhandledError" Nov 23 23:23:25.603295 kubelet[2297]: I1123 23:23:25.603281 2297 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 23 23:23:25.603416 kubelet[2297]: W1123 23:23:25.603402 2297 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Nov 23 23:23:25.604623 kubelet[2297]: I1123 23:23:25.604589 2297 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 23 23:23:25.604623 kubelet[2297]: I1123 23:23:25.604629 2297 server.go:1287] "Started kubelet" Nov 23 23:23:25.604834 kubelet[2297]: I1123 23:23:25.604799 2297 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 23 23:23:25.607109 kubelet[2297]: I1123 23:23:25.607085 2297 server.go:479] "Adding debug handlers to kubelet server" Nov 23 23:23:25.607452 kubelet[2297]: I1123 23:23:25.607394 2297 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 23 23:23:25.607779 kubelet[2297]: I1123 23:23:25.607750 2297 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 23 23:23:25.608211 kubelet[2297]: E1123 23:23:25.607911 2297 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.134:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.134:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.187ac64570178f8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-11-23 23:23:25.604605837 +0000 UTC m=+1.152957927,LastTimestamp:2025-11-23 23:23:25.604605837 +0000 UTC m=+1.152957927,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Nov 23 23:23:25.610605 kubelet[2297]: I1123 23:23:25.610584 2297 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 23 23:23:25.611560 kubelet[2297]: I1123 23:23:25.610598 2297 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 23 23:23:25.611560 kubelet[2297]: I1123 23:23:25.611552 2297 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 23 23:23:25.611658 kubelet[2297]: E1123 23:23:25.610875 2297 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 23 23:23:25.611658 kubelet[2297]: I1123 23:23:25.611648 2297 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 23 23:23:25.611889 kubelet[2297]: I1123 23:23:25.611861 2297 reconciler.go:26] "Reconciler: start to sync state" Nov 23 23:23:25.612166 kubelet[2297]: E1123 23:23:25.611992 2297 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 23 23:23:25.612166 kubelet[2297]: W1123 23:23:25.611995 2297 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.134:6443: connect: connection refused Nov 23 23:23:25.612166 kubelet[2297]: E1123 23:23:25.612053 2297 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.134:6443: connect: connection refused" logger="UnhandledError" Nov 23 23:23:25.613083 kubelet[2297]: E1123 23:23:25.613047 2297 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.134:6443: connect: connection refused" interval="200ms" Nov 23 23:23:25.617517 kubelet[2297]: I1123 23:23:25.616269 2297 factory.go:221] Registration of the containerd container factory successfully Nov 23 23:23:25.617517 kubelet[2297]: I1123 23:23:25.616291 2297 factory.go:221] Registration of the systemd container factory successfully Nov 23 23:23:25.617517 kubelet[2297]: I1123 23:23:25.616380 2297 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 23 23:23:25.626678 kubelet[2297]: I1123 23:23:25.626617 2297 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 23 23:23:25.626678 kubelet[2297]: I1123 23:23:25.626634 2297 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 23 23:23:25.626678 kubelet[2297]: I1123 23:23:25.626653 2297 state_mem.go:36] "Initialized new in-memory state store" Nov 23 23:23:25.628076 kubelet[2297]: I1123 23:23:25.627937 2297 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 23 23:23:25.629034 kubelet[2297]: I1123 23:23:25.629004 2297 policy_none.go:49] "None policy: Start" Nov 23 23:23:25.629034 kubelet[2297]: I1123 23:23:25.629028 2297 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 23 23:23:25.629034 kubelet[2297]: I1123 23:23:25.629039 2297 state_mem.go:35] "Initializing new in-memory state store" Nov 23 23:23:25.629212 kubelet[2297]: I1123 23:23:25.629195 2297 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 23 23:23:25.629267 kubelet[2297]: I1123 23:23:25.629260 2297 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 23 23:23:25.629501 kubelet[2297]: I1123 23:23:25.629475 2297 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 23 23:23:25.629608 kubelet[2297]: I1123 23:23:25.629543 2297 kubelet.go:2382] "Starting kubelet main sync loop" Nov 23 23:23:25.629691 kubelet[2297]: E1123 23:23:25.629673 2297 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 23 23:23:25.630311 kubelet[2297]: W1123 23:23:25.630281 2297 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.134:6443: connect: connection refused Nov 23 23:23:25.630496 kubelet[2297]: E1123 23:23:25.630325 2297 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.134:6443: connect: connection refused" logger="UnhandledError" Nov 23 23:23:25.635908 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Nov 23 23:23:25.649261 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Nov 23 23:23:25.652276 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Nov 23 23:23:25.668424 kubelet[2297]: I1123 23:23:25.668371 2297 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 23 23:23:25.669428 kubelet[2297]: I1123 23:23:25.669070 2297 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 23 23:23:25.669428 kubelet[2297]: I1123 23:23:25.669089 2297 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 23 23:23:25.669428 kubelet[2297]: I1123 23:23:25.669356 2297 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 23 23:23:25.670441 kubelet[2297]: E1123 23:23:25.670412 2297 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 23 23:23:25.670520 kubelet[2297]: E1123 23:23:25.670452 2297 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Nov 23 23:23:25.737298 systemd[1]: Created slice kubepods-burstable-podfa0a4945e2ec03c0c5171164bad4ce84.slice - libcontainer container kubepods-burstable-podfa0a4945e2ec03c0c5171164bad4ce84.slice. Nov 23 23:23:25.746183 kubelet[2297]: E1123 23:23:25.746130 2297 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 23 23:23:25.748232 systemd[1]: Created slice kubepods-burstable-pod55d9ac750f8c9141f337af8b08cf5c9d.slice - libcontainer container kubepods-burstable-pod55d9ac750f8c9141f337af8b08cf5c9d.slice. Nov 23 23:23:25.755605 kubelet[2297]: E1123 23:23:25.755539 2297 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 23 23:23:25.757785 systemd[1]: Created slice kubepods-burstable-pod0a68423804124305a9de061f38780871.slice - libcontainer container kubepods-burstable-pod0a68423804124305a9de061f38780871.slice. Nov 23 23:23:25.759261 kubelet[2297]: E1123 23:23:25.759216 2297 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 23 23:23:25.771513 kubelet[2297]: I1123 23:23:25.771343 2297 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 23 23:23:25.771886 kubelet[2297]: E1123 23:23:25.771848 2297 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.134:6443/api/v1/nodes\": dial tcp 10.0.0.134:6443: connect: connection refused" node="localhost" Nov 23 23:23:25.813360 kubelet[2297]: I1123 23:23:25.813236 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:25.813360 kubelet[2297]: I1123 23:23:25.813281 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:25.813360 kubelet[2297]: I1123 23:23:25.813309 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fa0a4945e2ec03c0c5171164bad4ce84-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"fa0a4945e2ec03c0c5171164bad4ce84\") " pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:25.813360 kubelet[2297]: I1123 23:23:25.813328 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fa0a4945e2ec03c0c5171164bad4ce84-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"fa0a4945e2ec03c0c5171164bad4ce84\") " pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:25.813360 kubelet[2297]: I1123 23:23:25.813351 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:25.813568 kubelet[2297]: I1123 23:23:25.813368 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:25.813568 kubelet[2297]: I1123 23:23:25.813384 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:25.813568 kubelet[2297]: I1123 23:23:25.813401 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0a68423804124305a9de061f38780871-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"0a68423804124305a9de061f38780871\") " pod="kube-system/kube-scheduler-localhost" Nov 23 23:23:25.813568 kubelet[2297]: I1123 23:23:25.813416 2297 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fa0a4945e2ec03c0c5171164bad4ce84-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"fa0a4945e2ec03c0c5171164bad4ce84\") " pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:25.815721 kubelet[2297]: E1123 23:23:25.815683 2297 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.134:6443: connect: connection refused" interval="400ms" Nov 23 23:23:25.973472 kubelet[2297]: I1123 23:23:25.973445 2297 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 23 23:23:25.973896 kubelet[2297]: E1123 23:23:25.973862 2297 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.134:6443/api/v1/nodes\": dial tcp 10.0.0.134:6443: connect: connection refused" node="localhost" Nov 23 23:23:26.048041 containerd[1527]: time="2025-11-23T23:23:26.047983352Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:fa0a4945e2ec03c0c5171164bad4ce84,Namespace:kube-system,Attempt:0,}" Nov 23 23:23:26.056827 containerd[1527]: time="2025-11-23T23:23:26.056743183Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:55d9ac750f8c9141f337af8b08cf5c9d,Namespace:kube-system,Attempt:0,}" Nov 23 23:23:26.061509 containerd[1527]: time="2025-11-23T23:23:26.061319097Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:0a68423804124305a9de061f38780871,Namespace:kube-system,Attempt:0,}" Nov 23 23:23:26.065537 containerd[1527]: time="2025-11-23T23:23:26.065418876Z" level=info msg="connecting to shim 6a5ff2a6b2b4a2ca1599aa780538be26630c1fc447c960abf9399b1c49452633" address="unix:///run/containerd/s/4033072d38e70cc16b754d3ce56d24836ec6b75528a743c37d43019ce6492571" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:23:26.085234 containerd[1527]: time="2025-11-23T23:23:26.085186225Z" level=info msg="connecting to shim cc8507d262067c9f33e2172e875ab7255a0b29a78412da1168ccfaed84fe44c2" address="unix:///run/containerd/s/202b42656fefb1e7b60d0ff8d14ae1a55bebdfe8402dfd58312735f2fd46fd1c" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:23:26.092718 containerd[1527]: time="2025-11-23T23:23:26.092670064Z" level=info msg="connecting to shim 5a88494b2c37c447c9533c1c8a6c6b8933c24686ce04022eeb03e7b246ef282b" address="unix:///run/containerd/s/b8ef5d6f05e74420034439c424736bcfecfbeadb9a4bc45b36ccdf07c88e6c1d" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:23:26.092720 systemd[1]: Started cri-containerd-6a5ff2a6b2b4a2ca1599aa780538be26630c1fc447c960abf9399b1c49452633.scope - libcontainer container 6a5ff2a6b2b4a2ca1599aa780538be26630c1fc447c960abf9399b1c49452633. Nov 23 23:23:26.115657 systemd[1]: Started cri-containerd-cc8507d262067c9f33e2172e875ab7255a0b29a78412da1168ccfaed84fe44c2.scope - libcontainer container cc8507d262067c9f33e2172e875ab7255a0b29a78412da1168ccfaed84fe44c2. Nov 23 23:23:26.120073 systemd[1]: Started cri-containerd-5a88494b2c37c447c9533c1c8a6c6b8933c24686ce04022eeb03e7b246ef282b.scope - libcontainer container 5a88494b2c37c447c9533c1c8a6c6b8933c24686ce04022eeb03e7b246ef282b. Nov 23 23:23:26.154139 containerd[1527]: time="2025-11-23T23:23:26.154096166Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:fa0a4945e2ec03c0c5171164bad4ce84,Namespace:kube-system,Attempt:0,} returns sandbox id \"6a5ff2a6b2b4a2ca1599aa780538be26630c1fc447c960abf9399b1c49452633\"" Nov 23 23:23:26.157312 containerd[1527]: time="2025-11-23T23:23:26.157271578Z" level=info msg="CreateContainer within sandbox \"6a5ff2a6b2b4a2ca1599aa780538be26630c1fc447c960abf9399b1c49452633\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Nov 23 23:23:26.158180 containerd[1527]: time="2025-11-23T23:23:26.158143664Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:55d9ac750f8c9141f337af8b08cf5c9d,Namespace:kube-system,Attempt:0,} returns sandbox id \"cc8507d262067c9f33e2172e875ab7255a0b29a78412da1168ccfaed84fe44c2\"" Nov 23 23:23:26.160849 containerd[1527]: time="2025-11-23T23:23:26.160817240Z" level=info msg="CreateContainer within sandbox \"cc8507d262067c9f33e2172e875ab7255a0b29a78412da1168ccfaed84fe44c2\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Nov 23 23:23:26.164888 containerd[1527]: time="2025-11-23T23:23:26.164855104Z" level=info msg="Container 8c278a1a66758e434a34463deda13104c0184da22c470918fc7fe9afca8bc3b9: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:23:26.166935 containerd[1527]: time="2025-11-23T23:23:26.166655196Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:0a68423804124305a9de061f38780871,Namespace:kube-system,Attempt:0,} returns sandbox id \"5a88494b2c37c447c9533c1c8a6c6b8933c24686ce04022eeb03e7b246ef282b\"" Nov 23 23:23:26.169333 containerd[1527]: time="2025-11-23T23:23:26.169302511Z" level=info msg="CreateContainer within sandbox \"5a88494b2c37c447c9533c1c8a6c6b8933c24686ce04022eeb03e7b246ef282b\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Nov 23 23:23:26.172375 containerd[1527]: time="2025-11-23T23:23:26.172333907Z" level=info msg="Container 18a36188ee9da75941d0be7d2f76558916ee1d325e32015d430fe32598225ba8: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:23:26.174965 containerd[1527]: time="2025-11-23T23:23:26.174930499Z" level=info msg="CreateContainer within sandbox \"6a5ff2a6b2b4a2ca1599aa780538be26630c1fc447c960abf9399b1c49452633\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"8c278a1a66758e434a34463deda13104c0184da22c470918fc7fe9afca8bc3b9\"" Nov 23 23:23:26.175752 containerd[1527]: time="2025-11-23T23:23:26.175724482Z" level=info msg="StartContainer for \"8c278a1a66758e434a34463deda13104c0184da22c470918fc7fe9afca8bc3b9\"" Nov 23 23:23:26.177081 containerd[1527]: time="2025-11-23T23:23:26.177045402Z" level=info msg="connecting to shim 8c278a1a66758e434a34463deda13104c0184da22c470918fc7fe9afca8bc3b9" address="unix:///run/containerd/s/4033072d38e70cc16b754d3ce56d24836ec6b75528a743c37d43019ce6492571" protocol=ttrpc version=3 Nov 23 23:23:26.178089 containerd[1527]: time="2025-11-23T23:23:26.178047353Z" level=info msg="Container 9879e35855c1760c709408515d25a898bb2e3a78f1b05943a4d91eb934e5f4c1: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:23:26.182795 containerd[1527]: time="2025-11-23T23:23:26.182704887Z" level=info msg="CreateContainer within sandbox \"cc8507d262067c9f33e2172e875ab7255a0b29a78412da1168ccfaed84fe44c2\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"18a36188ee9da75941d0be7d2f76558916ee1d325e32015d430fe32598225ba8\"" Nov 23 23:23:26.183597 containerd[1527]: time="2025-11-23T23:23:26.183562904Z" level=info msg="StartContainer for \"18a36188ee9da75941d0be7d2f76558916ee1d325e32015d430fe32598225ba8\"" Nov 23 23:23:26.184852 containerd[1527]: time="2025-11-23T23:23:26.184802562Z" level=info msg="connecting to shim 18a36188ee9da75941d0be7d2f76558916ee1d325e32015d430fe32598225ba8" address="unix:///run/containerd/s/202b42656fefb1e7b60d0ff8d14ae1a55bebdfe8402dfd58312735f2fd46fd1c" protocol=ttrpc version=3 Nov 23 23:23:26.191189 containerd[1527]: time="2025-11-23T23:23:26.191109177Z" level=info msg="CreateContainer within sandbox \"5a88494b2c37c447c9533c1c8a6c6b8933c24686ce04022eeb03e7b246ef282b\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"9879e35855c1760c709408515d25a898bb2e3a78f1b05943a4d91eb934e5f4c1\"" Nov 23 23:23:26.191790 containerd[1527]: time="2025-11-23T23:23:26.191764861Z" level=info msg="StartContainer for \"9879e35855c1760c709408515d25a898bb2e3a78f1b05943a4d91eb934e5f4c1\"" Nov 23 23:23:26.192960 containerd[1527]: time="2025-11-23T23:23:26.192927216Z" level=info msg="connecting to shim 9879e35855c1760c709408515d25a898bb2e3a78f1b05943a4d91eb934e5f4c1" address="unix:///run/containerd/s/b8ef5d6f05e74420034439c424736bcfecfbeadb9a4bc45b36ccdf07c88e6c1d" protocol=ttrpc version=3 Nov 23 23:23:26.206720 systemd[1]: Started cri-containerd-18a36188ee9da75941d0be7d2f76558916ee1d325e32015d430fe32598225ba8.scope - libcontainer container 18a36188ee9da75941d0be7d2f76558916ee1d325e32015d430fe32598225ba8. Nov 23 23:23:26.210298 systemd[1]: Started cri-containerd-8c278a1a66758e434a34463deda13104c0184da22c470918fc7fe9afca8bc3b9.scope - libcontainer container 8c278a1a66758e434a34463deda13104c0184da22c470918fc7fe9afca8bc3b9. Nov 23 23:23:26.216509 kubelet[2297]: E1123 23:23:26.216268 2297 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.134:6443: connect: connection refused" interval="800ms" Nov 23 23:23:26.223658 systemd[1]: Started cri-containerd-9879e35855c1760c709408515d25a898bb2e3a78f1b05943a4d91eb934e5f4c1.scope - libcontainer container 9879e35855c1760c709408515d25a898bb2e3a78f1b05943a4d91eb934e5f4c1. Nov 23 23:23:26.264506 containerd[1527]: time="2025-11-23T23:23:26.263414171Z" level=info msg="StartContainer for \"18a36188ee9da75941d0be7d2f76558916ee1d325e32015d430fe32598225ba8\" returns successfully" Nov 23 23:23:26.266198 containerd[1527]: time="2025-11-23T23:23:26.266143666Z" level=info msg="StartContainer for \"8c278a1a66758e434a34463deda13104c0184da22c470918fc7fe9afca8bc3b9\" returns successfully" Nov 23 23:23:26.272960 containerd[1527]: time="2025-11-23T23:23:26.272539896Z" level=info msg="StartContainer for \"9879e35855c1760c709408515d25a898bb2e3a78f1b05943a4d91eb934e5f4c1\" returns successfully" Nov 23 23:23:26.375952 kubelet[2297]: I1123 23:23:26.375836 2297 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 23 23:23:26.637772 kubelet[2297]: E1123 23:23:26.637277 2297 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 23 23:23:26.638939 kubelet[2297]: E1123 23:23:26.638918 2297 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 23 23:23:26.640176 kubelet[2297]: E1123 23:23:26.640156 2297 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 23 23:23:27.643658 kubelet[2297]: E1123 23:23:27.643217 2297 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 23 23:23:27.643658 kubelet[2297]: E1123 23:23:27.643532 2297 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 23 23:23:28.126994 kubelet[2297]: E1123 23:23:28.126960 2297 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Nov 23 23:23:28.198785 kubelet[2297]: I1123 23:23:28.198742 2297 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Nov 23 23:23:28.213498 kubelet[2297]: I1123 23:23:28.212085 2297 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:28.221234 kubelet[2297]: E1123 23:23:28.221195 2297 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:28.221234 kubelet[2297]: I1123 23:23:28.221224 2297 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Nov 23 23:23:28.223216 kubelet[2297]: E1123 23:23:28.223195 2297 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Nov 23 23:23:28.223470 kubelet[2297]: I1123 23:23:28.223296 2297 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:28.226524 kubelet[2297]: E1123 23:23:28.226492 2297 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:28.601046 kubelet[2297]: I1123 23:23:28.600816 2297 apiserver.go:52] "Watching apiserver" Nov 23 23:23:28.612427 kubelet[2297]: I1123 23:23:28.612392 2297 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 23 23:23:28.643857 kubelet[2297]: I1123 23:23:28.643805 2297 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:28.645948 kubelet[2297]: E1123 23:23:28.645904 2297 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:29.795795 kubelet[2297]: I1123 23:23:29.795617 2297 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:30.383357 systemd[1]: Reload requested from client PID 2570 ('systemctl') (unit session-7.scope)... Nov 23 23:23:30.383372 systemd[1]: Reloading... Nov 23 23:23:30.440507 zram_generator::config[2613]: No configuration found. Nov 23 23:23:30.603771 systemd[1]: Reloading finished in 220 ms. Nov 23 23:23:30.630155 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Nov 23 23:23:30.645517 systemd[1]: kubelet.service: Deactivated successfully. Nov 23 23:23:30.645808 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 23 23:23:30.645863 systemd[1]: kubelet.service: Consumed 1.532s CPU time, 129.6M memory peak. Nov 23 23:23:30.647428 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 23 23:23:30.791038 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 23 23:23:30.806829 (kubelet)[2655]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 23 23:23:30.844644 kubelet[2655]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 23:23:30.844644 kubelet[2655]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 23 23:23:30.844644 kubelet[2655]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 23:23:30.844967 kubelet[2655]: I1123 23:23:30.844703 2655 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 23 23:23:30.850029 kubelet[2655]: I1123 23:23:30.849991 2655 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 23 23:23:30.850029 kubelet[2655]: I1123 23:23:30.850020 2655 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 23 23:23:30.850278 kubelet[2655]: I1123 23:23:30.850250 2655 server.go:954] "Client rotation is on, will bootstrap in background" Nov 23 23:23:30.851487 kubelet[2655]: I1123 23:23:30.851464 2655 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 23 23:23:30.853503 kubelet[2655]: I1123 23:23:30.853471 2655 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 23 23:23:30.859514 kubelet[2655]: I1123 23:23:30.858601 2655 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 23 23:23:30.861105 kubelet[2655]: I1123 23:23:30.861086 2655 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 23 23:23:30.861304 kubelet[2655]: I1123 23:23:30.861281 2655 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 23 23:23:30.861449 kubelet[2655]: I1123 23:23:30.861306 2655 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 23 23:23:30.861560 kubelet[2655]: I1123 23:23:30.861458 2655 topology_manager.go:138] "Creating topology manager with none policy" Nov 23 23:23:30.861560 kubelet[2655]: I1123 23:23:30.861466 2655 container_manager_linux.go:304] "Creating device plugin manager" Nov 23 23:23:30.861560 kubelet[2655]: I1123 23:23:30.861524 2655 state_mem.go:36] "Initialized new in-memory state store" Nov 23 23:23:30.861673 kubelet[2655]: I1123 23:23:30.861654 2655 kubelet.go:446] "Attempting to sync node with API server" Nov 23 23:23:30.861673 kubelet[2655]: I1123 23:23:30.861669 2655 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 23 23:23:30.861718 kubelet[2655]: I1123 23:23:30.861688 2655 kubelet.go:352] "Adding apiserver pod source" Nov 23 23:23:30.861718 kubelet[2655]: I1123 23:23:30.861697 2655 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 23 23:23:30.862312 kubelet[2655]: I1123 23:23:30.862291 2655 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Nov 23 23:23:30.862821 kubelet[2655]: I1123 23:23:30.862787 2655 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 23 23:23:30.863203 kubelet[2655]: I1123 23:23:30.863163 2655 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 23 23:23:30.863203 kubelet[2655]: I1123 23:23:30.863193 2655 server.go:1287] "Started kubelet" Nov 23 23:23:30.863474 kubelet[2655]: I1123 23:23:30.863446 2655 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 23 23:23:30.864293 kubelet[2655]: I1123 23:23:30.864264 2655 server.go:479] "Adding debug handlers to kubelet server" Nov 23 23:23:30.865380 kubelet[2655]: I1123 23:23:30.865316 2655 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 23 23:23:30.865576 kubelet[2655]: I1123 23:23:30.865553 2655 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 23 23:23:30.866892 kubelet[2655]: I1123 23:23:30.866866 2655 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 23 23:23:30.868026 kubelet[2655]: I1123 23:23:30.867861 2655 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 23 23:23:30.868170 kubelet[2655]: I1123 23:23:30.866954 2655 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 23 23:23:30.868270 kubelet[2655]: I1123 23:23:30.868244 2655 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 23 23:23:30.868853 kubelet[2655]: I1123 23:23:30.868729 2655 reconciler.go:26] "Reconciler: start to sync state" Nov 23 23:23:30.872072 kubelet[2655]: E1123 23:23:30.872048 2655 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 23 23:23:30.879720 kubelet[2655]: I1123 23:23:30.879671 2655 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 23 23:23:30.880594 kubelet[2655]: I1123 23:23:30.880572 2655 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 23 23:23:30.880635 kubelet[2655]: I1123 23:23:30.880599 2655 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 23 23:23:30.880746 kubelet[2655]: I1123 23:23:30.880713 2655 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 23 23:23:30.880746 kubelet[2655]: I1123 23:23:30.880745 2655 kubelet.go:2382] "Starting kubelet main sync loop" Nov 23 23:23:30.880829 kubelet[2655]: E1123 23:23:30.880809 2655 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 23 23:23:30.888439 kubelet[2655]: E1123 23:23:30.887302 2655 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 23 23:23:30.894709 kubelet[2655]: I1123 23:23:30.894682 2655 factory.go:221] Registration of the containerd container factory successfully Nov 23 23:23:30.896755 kubelet[2655]: I1123 23:23:30.894704 2655 factory.go:221] Registration of the systemd container factory successfully Nov 23 23:23:30.896755 kubelet[2655]: I1123 23:23:30.896674 2655 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 23 23:23:30.938956 kubelet[2655]: I1123 23:23:30.938926 2655 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 23 23:23:30.938956 kubelet[2655]: I1123 23:23:30.938947 2655 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 23 23:23:30.938956 kubelet[2655]: I1123 23:23:30.938965 2655 state_mem.go:36] "Initialized new in-memory state store" Nov 23 23:23:30.940056 kubelet[2655]: I1123 23:23:30.940033 2655 state_mem.go:88] "Updated default CPUSet" cpuSet="" Nov 23 23:23:30.940094 kubelet[2655]: I1123 23:23:30.940057 2655 state_mem.go:96] "Updated CPUSet assignments" assignments={} Nov 23 23:23:30.940094 kubelet[2655]: I1123 23:23:30.940082 2655 policy_none.go:49] "None policy: Start" Nov 23 23:23:30.940094 kubelet[2655]: I1123 23:23:30.940090 2655 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 23 23:23:30.940166 kubelet[2655]: I1123 23:23:30.940103 2655 state_mem.go:35] "Initializing new in-memory state store" Nov 23 23:23:30.940407 kubelet[2655]: I1123 23:23:30.940392 2655 state_mem.go:75] "Updated machine memory state" Nov 23 23:23:30.946986 kubelet[2655]: I1123 23:23:30.946962 2655 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 23 23:23:30.947164 kubelet[2655]: I1123 23:23:30.947114 2655 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 23 23:23:30.947164 kubelet[2655]: I1123 23:23:30.947127 2655 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 23 23:23:30.947436 kubelet[2655]: I1123 23:23:30.947415 2655 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 23 23:23:30.949144 kubelet[2655]: E1123 23:23:30.949126 2655 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 23 23:23:30.982308 kubelet[2655]: I1123 23:23:30.982260 2655 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:30.982541 kubelet[2655]: I1123 23:23:30.982520 2655 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:30.982677 kubelet[2655]: I1123 23:23:30.982661 2655 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Nov 23 23:23:30.988770 kubelet[2655]: E1123 23:23:30.988731 2655 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:31.049252 kubelet[2655]: I1123 23:23:31.049220 2655 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 23 23:23:31.055199 kubelet[2655]: I1123 23:23:31.055166 2655 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Nov 23 23:23:31.055325 kubelet[2655]: I1123 23:23:31.055243 2655 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Nov 23 23:23:31.069736 kubelet[2655]: I1123 23:23:31.069706 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:31.069838 kubelet[2655]: I1123 23:23:31.069746 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:31.069838 kubelet[2655]: I1123 23:23:31.069770 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:31.069838 kubelet[2655]: I1123 23:23:31.069788 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0a68423804124305a9de061f38780871-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"0a68423804124305a9de061f38780871\") " pod="kube-system/kube-scheduler-localhost" Nov 23 23:23:31.069838 kubelet[2655]: I1123 23:23:31.069804 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fa0a4945e2ec03c0c5171164bad4ce84-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"fa0a4945e2ec03c0c5171164bad4ce84\") " pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:31.069838 kubelet[2655]: I1123 23:23:31.069819 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fa0a4945e2ec03c0c5171164bad4ce84-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"fa0a4945e2ec03c0c5171164bad4ce84\") " pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:31.069947 kubelet[2655]: I1123 23:23:31.069833 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fa0a4945e2ec03c0c5171164bad4ce84-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"fa0a4945e2ec03c0c5171164bad4ce84\") " pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:31.069947 kubelet[2655]: I1123 23:23:31.069865 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:31.069947 kubelet[2655]: I1123 23:23:31.069884 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/55d9ac750f8c9141f337af8b08cf5c9d-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"55d9ac750f8c9141f337af8b08cf5c9d\") " pod="kube-system/kube-controller-manager-localhost" Nov 23 23:23:31.862549 kubelet[2655]: I1123 23:23:31.862475 2655 apiserver.go:52] "Watching apiserver" Nov 23 23:23:31.868569 kubelet[2655]: I1123 23:23:31.868523 2655 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 23 23:23:31.917493 kubelet[2655]: I1123 23:23:31.917455 2655 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:31.926507 kubelet[2655]: E1123 23:23:31.924882 2655 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Nov 23 23:23:31.947281 kubelet[2655]: I1123 23:23:31.947213 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=2.947197637 podStartE2EDuration="2.947197637s" podCreationTimestamp="2025-11-23 23:23:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 23:23:31.947177844 +0000 UTC m=+1.137076269" watchObservedRunningTime="2025-11-23 23:23:31.947197637 +0000 UTC m=+1.137096022" Nov 23 23:23:31.969333 kubelet[2655]: I1123 23:23:31.968784 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.9687670769999999 podStartE2EDuration="1.968767077s" podCreationTimestamp="2025-11-23 23:23:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 23:23:31.96020159 +0000 UTC m=+1.150100015" watchObservedRunningTime="2025-11-23 23:23:31.968767077 +0000 UTC m=+1.158665462" Nov 23 23:23:31.980828 kubelet[2655]: I1123 23:23:31.980044 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.9800271189999998 podStartE2EDuration="1.980027119s" podCreationTimestamp="2025-11-23 23:23:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 23:23:31.969500963 +0000 UTC m=+1.159399428" watchObservedRunningTime="2025-11-23 23:23:31.980027119 +0000 UTC m=+1.169925544" Nov 23 23:23:35.558050 kubelet[2655]: I1123 23:23:35.558017 2655 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Nov 23 23:23:35.558790 containerd[1527]: time="2025-11-23T23:23:35.558750754Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Nov 23 23:23:35.559170 kubelet[2655]: I1123 23:23:35.559145 2655 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Nov 23 23:23:36.223883 systemd[1]: Created slice kubepods-besteffort-pod0438799c_5f4f_4a14_a605_13d315c0bc56.slice - libcontainer container kubepods-besteffort-pod0438799c_5f4f_4a14_a605_13d315c0bc56.slice. Nov 23 23:23:36.303931 kubelet[2655]: I1123 23:23:36.303885 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/0438799c-5f4f-4a14-a605-13d315c0bc56-kube-proxy\") pod \"kube-proxy-9pqcs\" (UID: \"0438799c-5f4f-4a14-a605-13d315c0bc56\") " pod="kube-system/kube-proxy-9pqcs" Nov 23 23:23:36.304399 kubelet[2655]: I1123 23:23:36.304348 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0438799c-5f4f-4a14-a605-13d315c0bc56-xtables-lock\") pod \"kube-proxy-9pqcs\" (UID: \"0438799c-5f4f-4a14-a605-13d315c0bc56\") " pod="kube-system/kube-proxy-9pqcs" Nov 23 23:23:36.304707 kubelet[2655]: I1123 23:23:36.304647 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzcfx\" (UniqueName: \"kubernetes.io/projected/0438799c-5f4f-4a14-a605-13d315c0bc56-kube-api-access-jzcfx\") pod \"kube-proxy-9pqcs\" (UID: \"0438799c-5f4f-4a14-a605-13d315c0bc56\") " pod="kube-system/kube-proxy-9pqcs" Nov 23 23:23:36.304812 kubelet[2655]: I1123 23:23:36.304797 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0438799c-5f4f-4a14-a605-13d315c0bc56-lib-modules\") pod \"kube-proxy-9pqcs\" (UID: \"0438799c-5f4f-4a14-a605-13d315c0bc56\") " pod="kube-system/kube-proxy-9pqcs" Nov 23 23:23:36.541669 containerd[1527]: time="2025-11-23T23:23:36.541568918Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-9pqcs,Uid:0438799c-5f4f-4a14-a605-13d315c0bc56,Namespace:kube-system,Attempt:0,}" Nov 23 23:23:36.556279 containerd[1527]: time="2025-11-23T23:23:36.555970224Z" level=info msg="connecting to shim 2f88be7083ad975c9f00f975ba102344eefd99f39a1907fbd115f4ea1e24d794" address="unix:///run/containerd/s/6f8c43af24f392b1b2f4a19e464e36ed58d5a893e6f5adbb0956d26ef0864b23" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:23:36.580635 systemd[1]: Started cri-containerd-2f88be7083ad975c9f00f975ba102344eefd99f39a1907fbd115f4ea1e24d794.scope - libcontainer container 2f88be7083ad975c9f00f975ba102344eefd99f39a1907fbd115f4ea1e24d794. Nov 23 23:23:36.601918 containerd[1527]: time="2025-11-23T23:23:36.601876194Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-9pqcs,Uid:0438799c-5f4f-4a14-a605-13d315c0bc56,Namespace:kube-system,Attempt:0,} returns sandbox id \"2f88be7083ad975c9f00f975ba102344eefd99f39a1907fbd115f4ea1e24d794\"" Nov 23 23:23:36.608883 containerd[1527]: time="2025-11-23T23:23:36.608847099Z" level=info msg="CreateContainer within sandbox \"2f88be7083ad975c9f00f975ba102344eefd99f39a1907fbd115f4ea1e24d794\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Nov 23 23:23:36.626476 containerd[1527]: time="2025-11-23T23:23:36.624630970Z" level=info msg="Container 12de80f4f841080c1e64de7afad10d441cf4ea84a68d8fc012b3fd136eaa2b67: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:23:36.632706 containerd[1527]: time="2025-11-23T23:23:36.632602181Z" level=info msg="CreateContainer within sandbox \"2f88be7083ad975c9f00f975ba102344eefd99f39a1907fbd115f4ea1e24d794\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"12de80f4f841080c1e64de7afad10d441cf4ea84a68d8fc012b3fd136eaa2b67\"" Nov 23 23:23:36.633890 containerd[1527]: time="2025-11-23T23:23:36.633849874Z" level=info msg="StartContainer for \"12de80f4f841080c1e64de7afad10d441cf4ea84a68d8fc012b3fd136eaa2b67\"" Nov 23 23:23:36.635526 containerd[1527]: time="2025-11-23T23:23:36.635464507Z" level=info msg="connecting to shim 12de80f4f841080c1e64de7afad10d441cf4ea84a68d8fc012b3fd136eaa2b67" address="unix:///run/containerd/s/6f8c43af24f392b1b2f4a19e464e36ed58d5a893e6f5adbb0956d26ef0864b23" protocol=ttrpc version=3 Nov 23 23:23:36.644152 systemd[1]: Created slice kubepods-besteffort-pod6336d6d2_375e_46b4_8216_69b5993e1b1a.slice - libcontainer container kubepods-besteffort-pod6336d6d2_375e_46b4_8216_69b5993e1b1a.slice. Nov 23 23:23:36.659659 systemd[1]: Started cri-containerd-12de80f4f841080c1e64de7afad10d441cf4ea84a68d8fc012b3fd136eaa2b67.scope - libcontainer container 12de80f4f841080c1e64de7afad10d441cf4ea84a68d8fc012b3fd136eaa2b67. Nov 23 23:23:36.707208 kubelet[2655]: I1123 23:23:36.707168 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/6336d6d2-375e-46b4-8216-69b5993e1b1a-var-lib-calico\") pod \"tigera-operator-7dcd859c48-tsl5p\" (UID: \"6336d6d2-375e-46b4-8216-69b5993e1b1a\") " pod="tigera-operator/tigera-operator-7dcd859c48-tsl5p" Nov 23 23:23:36.707640 kubelet[2655]: I1123 23:23:36.707583 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfsb5\" (UniqueName: \"kubernetes.io/projected/6336d6d2-375e-46b4-8216-69b5993e1b1a-kube-api-access-pfsb5\") pod \"tigera-operator-7dcd859c48-tsl5p\" (UID: \"6336d6d2-375e-46b4-8216-69b5993e1b1a\") " pod="tigera-operator/tigera-operator-7dcd859c48-tsl5p" Nov 23 23:23:36.744147 containerd[1527]: time="2025-11-23T23:23:36.744038025Z" level=info msg="StartContainer for \"12de80f4f841080c1e64de7afad10d441cf4ea84a68d8fc012b3fd136eaa2b67\" returns successfully" Nov 23 23:23:36.944122 kubelet[2655]: I1123 23:23:36.943769 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-9pqcs" podStartSLOduration=0.94375424 podStartE2EDuration="943.75424ms" podCreationTimestamp="2025-11-23 23:23:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 23:23:36.943641526 +0000 UTC m=+6.133539911" watchObservedRunningTime="2025-11-23 23:23:36.94375424 +0000 UTC m=+6.133652665" Nov 23 23:23:36.948703 containerd[1527]: time="2025-11-23T23:23:36.948664136Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-tsl5p,Uid:6336d6d2-375e-46b4-8216-69b5993e1b1a,Namespace:tigera-operator,Attempt:0,}" Nov 23 23:23:36.964001 containerd[1527]: time="2025-11-23T23:23:36.963963793Z" level=info msg="connecting to shim 72a66cb2abfbd685452581d03e58191f54e90da83fc42fa49484998558e6b1b8" address="unix:///run/containerd/s/cf67d9a1188a1c65dc43f718719f0bf8f93083e6b12967b9bc1b4f9b154585fc" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:23:36.984660 systemd[1]: Started cri-containerd-72a66cb2abfbd685452581d03e58191f54e90da83fc42fa49484998558e6b1b8.scope - libcontainer container 72a66cb2abfbd685452581d03e58191f54e90da83fc42fa49484998558e6b1b8. Nov 23 23:23:37.019362 containerd[1527]: time="2025-11-23T23:23:37.019325190Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-tsl5p,Uid:6336d6d2-375e-46b4-8216-69b5993e1b1a,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"72a66cb2abfbd685452581d03e58191f54e90da83fc42fa49484998558e6b1b8\"" Nov 23 23:23:37.022109 containerd[1527]: time="2025-11-23T23:23:37.022078050Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Nov 23 23:23:38.402962 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1495057758.mount: Deactivated successfully. Nov 23 23:23:39.270974 containerd[1527]: time="2025-11-23T23:23:39.270913003Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:39.271500 containerd[1527]: time="2025-11-23T23:23:39.271449498Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=22152004" Nov 23 23:23:39.272226 containerd[1527]: time="2025-11-23T23:23:39.272201504Z" level=info msg="ImageCreate event name:\"sha256:19f52e4b7ea471a91d4186e9701288b905145dc20d4928cbbf2eac8d9dfce54b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:39.274326 containerd[1527]: time="2025-11-23T23:23:39.274294569Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:39.274925 containerd[1527]: time="2025-11-23T23:23:39.274898342Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:19f52e4b7ea471a91d4186e9701288b905145dc20d4928cbbf2eac8d9dfce54b\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"22147999\" in 2.252787653s" Nov 23 23:23:39.275117 containerd[1527]: time="2025-11-23T23:23:39.274930460Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:19f52e4b7ea471a91d4186e9701288b905145dc20d4928cbbf2eac8d9dfce54b\"" Nov 23 23:23:39.276738 containerd[1527]: time="2025-11-23T23:23:39.276709059Z" level=info msg="CreateContainer within sandbox \"72a66cb2abfbd685452581d03e58191f54e90da83fc42fa49484998558e6b1b8\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Nov 23 23:23:39.284359 containerd[1527]: time="2025-11-23T23:23:39.283804337Z" level=info msg="Container 18a6e149126058405c55f6d130b3f39379f1136de8647eff150313a259959628: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:23:39.290161 containerd[1527]: time="2025-11-23T23:23:39.290131329Z" level=info msg="CreateContainer within sandbox \"72a66cb2abfbd685452581d03e58191f54e90da83fc42fa49484998558e6b1b8\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"18a6e149126058405c55f6d130b3f39379f1136de8647eff150313a259959628\"" Nov 23 23:23:39.291564 containerd[1527]: time="2025-11-23T23:23:39.291237079Z" level=info msg="StartContainer for \"18a6e149126058405c55f6d130b3f39379f1136de8647eff150313a259959628\"" Nov 23 23:23:39.292236 containerd[1527]: time="2025-11-23T23:23:39.292204435Z" level=info msg="connecting to shim 18a6e149126058405c55f6d130b3f39379f1136de8647eff150313a259959628" address="unix:///run/containerd/s/cf67d9a1188a1c65dc43f718719f0bf8f93083e6b12967b9bc1b4f9b154585fc" protocol=ttrpc version=3 Nov 23 23:23:39.311638 systemd[1]: Started cri-containerd-18a6e149126058405c55f6d130b3f39379f1136de8647eff150313a259959628.scope - libcontainer container 18a6e149126058405c55f6d130b3f39379f1136de8647eff150313a259959628. Nov 23 23:23:39.336714 containerd[1527]: time="2025-11-23T23:23:39.336569099Z" level=info msg="StartContainer for \"18a6e149126058405c55f6d130b3f39379f1136de8647eff150313a259959628\" returns successfully" Nov 23 23:23:43.431852 kubelet[2655]: I1123 23:23:43.431685 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-tsl5p" podStartSLOduration=5.177483903 podStartE2EDuration="7.431670688s" podCreationTimestamp="2025-11-23 23:23:36 +0000 UTC" firstStartedPulling="2025-11-23 23:23:37.021366047 +0000 UTC m=+6.211264472" lastFinishedPulling="2025-11-23 23:23:39.275552832 +0000 UTC m=+8.465451257" observedRunningTime="2025-11-23 23:23:39.947945515 +0000 UTC m=+9.137843940" watchObservedRunningTime="2025-11-23 23:23:43.431670688 +0000 UTC m=+12.621569113" Nov 23 23:23:44.632567 sudo[1741]: pam_unix(sudo:session): session closed for user root Nov 23 23:23:44.637304 sshd[1740]: Connection closed by 10.0.0.1 port 60732 Nov 23 23:23:44.637808 sshd-session[1736]: pam_unix(sshd:session): session closed for user core Nov 23 23:23:44.640783 systemd[1]: sshd@6-10.0.0.134:22-10.0.0.1:60732.service: Deactivated successfully. Nov 23 23:23:44.646063 systemd[1]: session-7.scope: Deactivated successfully. Nov 23 23:23:44.647069 systemd[1]: session-7.scope: Consumed 5.868s CPU time, 216.8M memory peak. Nov 23 23:23:44.648765 systemd-logind[1513]: Session 7 logged out. Waiting for processes to exit. Nov 23 23:23:44.651717 systemd-logind[1513]: Removed session 7. Nov 23 23:23:47.447018 update_engine[1514]: I20251123 23:23:47.446949 1514 update_attempter.cc:509] Updating boot flags... Nov 23 23:23:51.484428 systemd[1]: Created slice kubepods-besteffort-pod13e649b6_a65e_4e03_8afd_f55c6c1835a6.slice - libcontainer container kubepods-besteffort-pod13e649b6_a65e_4e03_8afd_f55c6c1835a6.slice. Nov 23 23:23:51.507333 kubelet[2655]: I1123 23:23:51.507235 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13e649b6-a65e-4e03-8afd-f55c6c1835a6-tigera-ca-bundle\") pod \"calico-typha-785cd77854-rh2sw\" (UID: \"13e649b6-a65e-4e03-8afd-f55c6c1835a6\") " pod="calico-system/calico-typha-785cd77854-rh2sw" Nov 23 23:23:51.507333 kubelet[2655]: I1123 23:23:51.507282 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/13e649b6-a65e-4e03-8afd-f55c6c1835a6-typha-certs\") pod \"calico-typha-785cd77854-rh2sw\" (UID: \"13e649b6-a65e-4e03-8afd-f55c6c1835a6\") " pod="calico-system/calico-typha-785cd77854-rh2sw" Nov 23 23:23:51.507333 kubelet[2655]: I1123 23:23:51.507300 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrrh9\" (UniqueName: \"kubernetes.io/projected/13e649b6-a65e-4e03-8afd-f55c6c1835a6-kube-api-access-jrrh9\") pod \"calico-typha-785cd77854-rh2sw\" (UID: \"13e649b6-a65e-4e03-8afd-f55c6c1835a6\") " pod="calico-system/calico-typha-785cd77854-rh2sw" Nov 23 23:23:51.652255 systemd[1]: Created slice kubepods-besteffort-podcceabd95_3b16_44c5_a9de_996c4e244a2a.slice - libcontainer container kubepods-besteffort-podcceabd95_3b16_44c5_a9de_996c4e244a2a.slice. Nov 23 23:23:51.708736 kubelet[2655]: I1123 23:23:51.708680 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/cceabd95-3b16-44c5-a9de-996c4e244a2a-node-certs\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.708872 kubelet[2655]: I1123 23:23:51.708767 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-cni-bin-dir\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.708872 kubelet[2655]: I1123 23:23:51.708787 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-cni-net-dir\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.708872 kubelet[2655]: I1123 23:23:51.708801 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-var-lib-calico\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.708872 kubelet[2655]: I1123 23:23:51.708865 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-policysync\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.712765 kubelet[2655]: I1123 23:23:51.708882 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-cni-log-dir\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.713037 kubelet[2655]: I1123 23:23:51.713012 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-flexvol-driver-host\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.713141 kubelet[2655]: I1123 23:23:51.713051 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cceabd95-3b16-44c5-a9de-996c4e244a2a-tigera-ca-bundle\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.713141 kubelet[2655]: I1123 23:23:51.713070 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-var-run-calico\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.713141 kubelet[2655]: I1123 23:23:51.713092 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-xtables-lock\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.713141 kubelet[2655]: I1123 23:23:51.713114 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cceabd95-3b16-44c5-a9de-996c4e244a2a-lib-modules\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.713247 kubelet[2655]: I1123 23:23:51.713148 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqq5b\" (UniqueName: \"kubernetes.io/projected/cceabd95-3b16-44c5-a9de-996c4e244a2a-kube-api-access-dqq5b\") pod \"calico-node-g2qw5\" (UID: \"cceabd95-3b16-44c5-a9de-996c4e244a2a\") " pod="calico-system/calico-node-g2qw5" Nov 23 23:23:51.792140 containerd[1527]: time="2025-11-23T23:23:51.792034472Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-785cd77854-rh2sw,Uid:13e649b6-a65e-4e03-8afd-f55c6c1835a6,Namespace:calico-system,Attempt:0,}" Nov 23 23:23:51.824728 kubelet[2655]: E1123 23:23:51.824595 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.824728 kubelet[2655]: W1123 23:23:51.824624 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.824728 kubelet[2655]: E1123 23:23:51.824652 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.829767 kubelet[2655]: E1123 23:23:51.829669 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.829767 kubelet[2655]: W1123 23:23:51.829691 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.829767 kubelet[2655]: E1123 23:23:51.829711 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.831412 kubelet[2655]: E1123 23:23:51.831351 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.831412 kubelet[2655]: W1123 23:23:51.831368 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.831412 kubelet[2655]: E1123 23:23:51.831381 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.833952 containerd[1527]: time="2025-11-23T23:23:51.833901972Z" level=info msg="connecting to shim 8332a67bc5e443399ca9493ceb6da3e9ad1362bd8d35daaa635a6e623b52695f" address="unix:///run/containerd/s/f2f009d741a39f74a51df79256f0de954aa950b0d618f02bd811c32c9f2ce414" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:23:51.857412 kubelet[2655]: E1123 23:23:51.857183 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:23:51.899699 systemd[1]: Started cri-containerd-8332a67bc5e443399ca9493ceb6da3e9ad1362bd8d35daaa635a6e623b52695f.scope - libcontainer container 8332a67bc5e443399ca9493ceb6da3e9ad1362bd8d35daaa635a6e623b52695f. Nov 23 23:23:51.905902 kubelet[2655]: E1123 23:23:51.905811 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.905902 kubelet[2655]: W1123 23:23:51.905835 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.905902 kubelet[2655]: E1123 23:23:51.905858 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.906311 kubelet[2655]: E1123 23:23:51.906221 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.906311 kubelet[2655]: W1123 23:23:51.906234 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.906311 kubelet[2655]: E1123 23:23:51.906276 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.906610 kubelet[2655]: E1123 23:23:51.906598 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.906675 kubelet[2655]: W1123 23:23:51.906663 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.906746 kubelet[2655]: E1123 23:23:51.906735 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.907095 kubelet[2655]: E1123 23:23:51.907021 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.907095 kubelet[2655]: W1123 23:23:51.907035 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.907095 kubelet[2655]: E1123 23:23:51.907046 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.907327 kubelet[2655]: E1123 23:23:51.907316 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.907428 kubelet[2655]: W1123 23:23:51.907379 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.907428 kubelet[2655]: E1123 23:23:51.907394 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.907673 kubelet[2655]: E1123 23:23:51.907661 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.907792 kubelet[2655]: W1123 23:23:51.907736 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.907792 kubelet[2655]: E1123 23:23:51.907752 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.907991 kubelet[2655]: E1123 23:23:51.907980 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.908149 kubelet[2655]: W1123 23:23:51.908044 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.908149 kubelet[2655]: E1123 23:23:51.908058 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.908332 kubelet[2655]: E1123 23:23:51.908267 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.908332 kubelet[2655]: W1123 23:23:51.908277 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.908332 kubelet[2655]: E1123 23:23:51.908286 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.908609 kubelet[2655]: E1123 23:23:51.908598 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.908748 kubelet[2655]: W1123 23:23:51.908668 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.908748 kubelet[2655]: E1123 23:23:51.908683 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.908981 kubelet[2655]: E1123 23:23:51.908969 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.909048 kubelet[2655]: W1123 23:23:51.909037 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.909108 kubelet[2655]: E1123 23:23:51.909097 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.909300 kubelet[2655]: E1123 23:23:51.909289 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.909357 kubelet[2655]: W1123 23:23:51.909347 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.909422 kubelet[2655]: E1123 23:23:51.909411 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.909716 kubelet[2655]: E1123 23:23:51.909651 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.909716 kubelet[2655]: W1123 23:23:51.909663 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.909716 kubelet[2655]: E1123 23:23:51.909674 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.909960 kubelet[2655]: E1123 23:23:51.909948 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.910106 kubelet[2655]: W1123 23:23:51.910014 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.910106 kubelet[2655]: E1123 23:23:51.910029 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.910244 kubelet[2655]: E1123 23:23:51.910234 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.910299 kubelet[2655]: W1123 23:23:51.910288 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.910352 kubelet[2655]: E1123 23:23:51.910342 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.910663 kubelet[2655]: E1123 23:23:51.910599 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.910663 kubelet[2655]: W1123 23:23:51.910612 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.910663 kubelet[2655]: E1123 23:23:51.910621 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.910923 kubelet[2655]: E1123 23:23:51.910911 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.911066 kubelet[2655]: W1123 23:23:51.910977 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.911066 kubelet[2655]: E1123 23:23:51.910989 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.911287 kubelet[2655]: E1123 23:23:51.911274 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.911449 kubelet[2655]: W1123 23:23:51.911355 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.911449 kubelet[2655]: E1123 23:23:51.911374 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.911622 kubelet[2655]: E1123 23:23:51.911611 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.911677 kubelet[2655]: W1123 23:23:51.911667 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.911723 kubelet[2655]: E1123 23:23:51.911713 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.912000 kubelet[2655]: E1123 23:23:51.911946 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.912000 kubelet[2655]: W1123 23:23:51.911957 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.912000 kubelet[2655]: E1123 23:23:51.911965 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.912335 kubelet[2655]: E1123 23:23:51.912321 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.912542 kubelet[2655]: W1123 23:23:51.912513 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.912605 kubelet[2655]: E1123 23:23:51.912593 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.914729 kubelet[2655]: E1123 23:23:51.914622 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.914729 kubelet[2655]: W1123 23:23:51.914638 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.914729 kubelet[2655]: E1123 23:23:51.914648 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.914729 kubelet[2655]: I1123 23:23:51.914674 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/82559865-90d3-4a8e-83a3-002c0c14990f-kubelet-dir\") pod \"csi-node-driver-b2nkd\" (UID: \"82559865-90d3-4a8e-83a3-002c0c14990f\") " pod="calico-system/csi-node-driver-b2nkd" Nov 23 23:23:51.915042 kubelet[2655]: E1123 23:23:51.915018 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.915129 kubelet[2655]: W1123 23:23:51.915102 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.915266 kubelet[2655]: E1123 23:23:51.915178 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.915266 kubelet[2655]: I1123 23:23:51.915216 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjff2\" (UniqueName: \"kubernetes.io/projected/82559865-90d3-4a8e-83a3-002c0c14990f-kube-api-access-bjff2\") pod \"csi-node-driver-b2nkd\" (UID: \"82559865-90d3-4a8e-83a3-002c0c14990f\") " pod="calico-system/csi-node-driver-b2nkd" Nov 23 23:23:51.915563 kubelet[2655]: E1123 23:23:51.915548 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.915678 kubelet[2655]: W1123 23:23:51.915663 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.915826 kubelet[2655]: E1123 23:23:51.915712 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.915826 kubelet[2655]: I1123 23:23:51.915731 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/82559865-90d3-4a8e-83a3-002c0c14990f-socket-dir\") pod \"csi-node-driver-b2nkd\" (UID: \"82559865-90d3-4a8e-83a3-002c0c14990f\") " pod="calico-system/csi-node-driver-b2nkd" Nov 23 23:23:51.916075 kubelet[2655]: E1123 23:23:51.916060 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.916167 kubelet[2655]: W1123 23:23:51.916154 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.916396 kubelet[2655]: E1123 23:23:51.916300 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.916396 kubelet[2655]: I1123 23:23:51.916330 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/82559865-90d3-4a8e-83a3-002c0c14990f-varrun\") pod \"csi-node-driver-b2nkd\" (UID: \"82559865-90d3-4a8e-83a3-002c0c14990f\") " pod="calico-system/csi-node-driver-b2nkd" Nov 23 23:23:51.916540 kubelet[2655]: E1123 23:23:51.916528 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.916605 kubelet[2655]: W1123 23:23:51.916594 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.916690 kubelet[2655]: E1123 23:23:51.916666 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.916894 kubelet[2655]: E1123 23:23:51.916882 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.916952 kubelet[2655]: W1123 23:23:51.916941 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.917021 kubelet[2655]: E1123 23:23:51.917003 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.917283 kubelet[2655]: E1123 23:23:51.917209 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.917283 kubelet[2655]: W1123 23:23:51.917237 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.917283 kubelet[2655]: E1123 23:23:51.917260 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.917677 kubelet[2655]: E1123 23:23:51.917579 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.917677 kubelet[2655]: W1123 23:23:51.917592 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.917677 kubelet[2655]: E1123 23:23:51.917609 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.917677 kubelet[2655]: I1123 23:23:51.917627 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/82559865-90d3-4a8e-83a3-002c0c14990f-registration-dir\") pod \"csi-node-driver-b2nkd\" (UID: \"82559865-90d3-4a8e-83a3-002c0c14990f\") " pod="calico-system/csi-node-driver-b2nkd" Nov 23 23:23:51.917946 kubelet[2655]: E1123 23:23:51.917933 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.918007 kubelet[2655]: W1123 23:23:51.917995 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.918080 kubelet[2655]: E1123 23:23:51.918068 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.918274 kubelet[2655]: E1123 23:23:51.918262 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.918274 kubelet[2655]: W1123 23:23:51.918320 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.918274 kubelet[2655]: E1123 23:23:51.918333 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.918650 kubelet[2655]: E1123 23:23:51.918638 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.918702 kubelet[2655]: W1123 23:23:51.918691 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.918769 kubelet[2655]: E1123 23:23:51.918758 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.918970 kubelet[2655]: E1123 23:23:51.918959 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.919030 kubelet[2655]: W1123 23:23:51.919019 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.919078 kubelet[2655]: E1123 23:23:51.919067 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.919286 kubelet[2655]: E1123 23:23:51.919275 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.919359 kubelet[2655]: W1123 23:23:51.919347 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.919424 kubelet[2655]: E1123 23:23:51.919413 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.919678 kubelet[2655]: E1123 23:23:51.919666 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.919813 kubelet[2655]: W1123 23:23:51.919746 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.919813 kubelet[2655]: E1123 23:23:51.919763 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.920153 kubelet[2655]: E1123 23:23:51.920126 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:51.920285 kubelet[2655]: W1123 23:23:51.920211 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:51.920285 kubelet[2655]: E1123 23:23:51.920244 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:51.950545 containerd[1527]: time="2025-11-23T23:23:51.950476933Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-785cd77854-rh2sw,Uid:13e649b6-a65e-4e03-8afd-f55c6c1835a6,Namespace:calico-system,Attempt:0,} returns sandbox id \"8332a67bc5e443399ca9493ceb6da3e9ad1362bd8d35daaa635a6e623b52695f\"" Nov 23 23:23:51.955430 containerd[1527]: time="2025-11-23T23:23:51.955396533Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Nov 23 23:23:51.957232 containerd[1527]: time="2025-11-23T23:23:51.957192849Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-g2qw5,Uid:cceabd95-3b16-44c5-a9de-996c4e244a2a,Namespace:calico-system,Attempt:0,}" Nov 23 23:23:52.020242 kubelet[2655]: E1123 23:23:52.019136 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020242 kubelet[2655]: W1123 23:23:52.019158 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020242 kubelet[2655]: E1123 23:23:52.019176 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020242 kubelet[2655]: E1123 23:23:52.019385 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020242 kubelet[2655]: W1123 23:23:52.019392 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020242 kubelet[2655]: E1123 23:23:52.019402 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020242 kubelet[2655]: E1123 23:23:52.019582 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020242 kubelet[2655]: W1123 23:23:52.019591 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020242 kubelet[2655]: E1123 23:23:52.019600 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020242 kubelet[2655]: E1123 23:23:52.019814 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020594 kubelet[2655]: W1123 23:23:52.019821 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020594 kubelet[2655]: E1123 23:23:52.019830 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020594 kubelet[2655]: E1123 23:23:52.019968 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020594 kubelet[2655]: W1123 23:23:52.019982 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020594 kubelet[2655]: E1123 23:23:52.019991 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020594 kubelet[2655]: E1123 23:23:52.020164 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020594 kubelet[2655]: W1123 23:23:52.020171 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020594 kubelet[2655]: E1123 23:23:52.020180 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020594 kubelet[2655]: E1123 23:23:52.020292 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020594 kubelet[2655]: W1123 23:23:52.020299 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020798 kubelet[2655]: E1123 23:23:52.020306 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020798 kubelet[2655]: E1123 23:23:52.020431 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020798 kubelet[2655]: W1123 23:23:52.020437 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020798 kubelet[2655]: E1123 23:23:52.020499 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020798 kubelet[2655]: E1123 23:23:52.020618 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020798 kubelet[2655]: W1123 23:23:52.020628 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020798 kubelet[2655]: E1123 23:23:52.020692 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.020926 kubelet[2655]: E1123 23:23:52.020839 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.020926 kubelet[2655]: W1123 23:23:52.020847 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.020998 kubelet[2655]: E1123 23:23:52.020975 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.021177 kubelet[2655]: E1123 23:23:52.021140 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.021177 kubelet[2655]: W1123 23:23:52.021174 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.021387 kubelet[2655]: E1123 23:23:52.021210 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.021538 kubelet[2655]: E1123 23:23:52.021336 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.021538 kubelet[2655]: W1123 23:23:52.021419 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.021538 kubelet[2655]: E1123 23:23:52.021470 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.022463 kubelet[2655]: E1123 23:23:52.022423 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.022463 kubelet[2655]: W1123 23:23:52.022438 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.022463 kubelet[2655]: E1123 23:23:52.022454 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.022684 kubelet[2655]: E1123 23:23:52.022642 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.022684 kubelet[2655]: W1123 23:23:52.022654 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.022684 kubelet[2655]: E1123 23:23:52.022666 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.022684 kubelet[2655]: E1123 23:23:52.022798 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.022684 kubelet[2655]: W1123 23:23:52.022804 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.022684 kubelet[2655]: E1123 23:23:52.022834 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.023918 kubelet[2655]: E1123 23:23:52.022961 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.023918 kubelet[2655]: W1123 23:23:52.022969 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.023918 kubelet[2655]: E1123 23:23:52.022990 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.023918 kubelet[2655]: E1123 23:23:52.023079 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.023918 kubelet[2655]: W1123 23:23:52.023086 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.023918 kubelet[2655]: E1123 23:23:52.023201 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.023918 kubelet[2655]: W1123 23:23:52.023207 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.023918 kubelet[2655]: E1123 23:23:52.023335 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.023918 kubelet[2655]: W1123 23:23:52.023342 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.023918 kubelet[2655]: E1123 23:23:52.023350 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.024126 kubelet[2655]: E1123 23:23:52.023535 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.024126 kubelet[2655]: E1123 23:23:52.023561 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.024126 kubelet[2655]: E1123 23:23:52.023717 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.024126 kubelet[2655]: W1123 23:23:52.023729 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.024126 kubelet[2655]: E1123 23:23:52.023744 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.024126 kubelet[2655]: E1123 23:23:52.024000 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.024126 kubelet[2655]: W1123 23:23:52.024011 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.024126 kubelet[2655]: E1123 23:23:52.024029 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.025441 kubelet[2655]: E1123 23:23:52.024232 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.025441 kubelet[2655]: W1123 23:23:52.024241 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.025441 kubelet[2655]: E1123 23:23:52.024255 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.025441 kubelet[2655]: E1123 23:23:52.024417 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.025441 kubelet[2655]: W1123 23:23:52.024425 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.025441 kubelet[2655]: E1123 23:23:52.024434 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.025441 kubelet[2655]: E1123 23:23:52.024626 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.025441 kubelet[2655]: W1123 23:23:52.024635 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.025441 kubelet[2655]: E1123 23:23:52.024644 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.025441 kubelet[2655]: E1123 23:23:52.024809 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.025714 kubelet[2655]: W1123 23:23:52.024816 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.025714 kubelet[2655]: E1123 23:23:52.024824 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.038666 kubelet[2655]: E1123 23:23:52.038630 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:52.038846 kubelet[2655]: W1123 23:23:52.038652 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:52.038846 kubelet[2655]: E1123 23:23:52.038814 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:52.038930 containerd[1527]: time="2025-11-23T23:23:52.038806983Z" level=info msg="connecting to shim 63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638" address="unix:///run/containerd/s/42d2b95fbb2b2e6d11cf4073a2f15e963ce0fdbf425861dfae92df00910752f4" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:23:52.070717 systemd[1]: Started cri-containerd-63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638.scope - libcontainer container 63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638. Nov 23 23:23:52.093808 containerd[1527]: time="2025-11-23T23:23:52.093752227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-g2qw5,Uid:cceabd95-3b16-44c5-a9de-996c4e244a2a,Namespace:calico-system,Attempt:0,} returns sandbox id \"63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638\"" Nov 23 23:23:52.892359 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount66973950.mount: Deactivated successfully. Nov 23 23:23:53.881899 kubelet[2655]: E1123 23:23:53.881858 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:23:53.926425 containerd[1527]: time="2025-11-23T23:23:53.926381667Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:53.927179 containerd[1527]: time="2025-11-23T23:23:53.927137770Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=33090687" Nov 23 23:23:53.927851 containerd[1527]: time="2025-11-23T23:23:53.927823915Z" level=info msg="ImageCreate event name:\"sha256:5fe38d12a54098df5aaf5ec7228dc2f976f60cb4f434d7256f03126b004fdc5b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:53.929524 containerd[1527]: time="2025-11-23T23:23:53.929496518Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:53.930215 containerd[1527]: time="2025-11-23T23:23:53.930191583Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:5fe38d12a54098df5aaf5ec7228dc2f976f60cb4f434d7256f03126b004fdc5b\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"33090541\" in 1.974560416s" Nov 23 23:23:53.930409 containerd[1527]: time="2025-11-23T23:23:53.930391658Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:5fe38d12a54098df5aaf5ec7228dc2f976f60cb4f434d7256f03126b004fdc5b\"" Nov 23 23:23:53.931321 containerd[1527]: time="2025-11-23T23:23:53.931298158Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Nov 23 23:23:53.960996 containerd[1527]: time="2025-11-23T23:23:53.960960700Z" level=info msg="CreateContainer within sandbox \"8332a67bc5e443399ca9493ceb6da3e9ad1362bd8d35daaa635a6e623b52695f\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Nov 23 23:23:53.967939 containerd[1527]: time="2025-11-23T23:23:53.967688511Z" level=info msg="Container 8ea3a3871eca94ba31fbb383569663e8a42ef755a380fa84f288a3e5f377518c: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:23:53.976386 containerd[1527]: time="2025-11-23T23:23:53.976209802Z" level=info msg="CreateContainer within sandbox \"8332a67bc5e443399ca9493ceb6da3e9ad1362bd8d35daaa635a6e623b52695f\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"8ea3a3871eca94ba31fbb383569663e8a42ef755a380fa84f288a3e5f377518c\"" Nov 23 23:23:53.978419 containerd[1527]: time="2025-11-23T23:23:53.978389674Z" level=info msg="StartContainer for \"8ea3a3871eca94ba31fbb383569663e8a42ef755a380fa84f288a3e5f377518c\"" Nov 23 23:23:53.979588 containerd[1527]: time="2025-11-23T23:23:53.979548488Z" level=info msg="connecting to shim 8ea3a3871eca94ba31fbb383569663e8a42ef755a380fa84f288a3e5f377518c" address="unix:///run/containerd/s/f2f009d741a39f74a51df79256f0de954aa950b0d618f02bd811c32c9f2ce414" protocol=ttrpc version=3 Nov 23 23:23:53.998652 systemd[1]: Started cri-containerd-8ea3a3871eca94ba31fbb383569663e8a42ef755a380fa84f288a3e5f377518c.scope - libcontainer container 8ea3a3871eca94ba31fbb383569663e8a42ef755a380fa84f288a3e5f377518c. Nov 23 23:23:54.040616 containerd[1527]: time="2025-11-23T23:23:54.040353938Z" level=info msg="StartContainer for \"8ea3a3871eca94ba31fbb383569663e8a42ef755a380fa84f288a3e5f377518c\" returns successfully" Nov 23 23:23:54.999726 kubelet[2655]: I1123 23:23:54.999656 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-785cd77854-rh2sw" podStartSLOduration=2.021157493 podStartE2EDuration="3.99941806s" podCreationTimestamp="2025-11-23 23:23:51 +0000 UTC" firstStartedPulling="2025-11-23 23:23:51.952753277 +0000 UTC m=+21.142651702" lastFinishedPulling="2025-11-23 23:23:53.931013884 +0000 UTC m=+23.120912269" observedRunningTime="2025-11-23 23:23:54.999159425 +0000 UTC m=+24.189057850" watchObservedRunningTime="2025-11-23 23:23:54.99941806 +0000 UTC m=+24.189316485" Nov 23 23:23:55.031438 kubelet[2655]: E1123 23:23:55.031396 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.031591 kubelet[2655]: W1123 23:23:55.031524 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.031591 kubelet[2655]: E1123 23:23:55.031569 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.031935 kubelet[2655]: E1123 23:23:55.031912 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.031964 kubelet[2655]: W1123 23:23:55.031935 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.031964 kubelet[2655]: E1123 23:23:55.031947 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.032324 kubelet[2655]: E1123 23:23:55.032246 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.032357 kubelet[2655]: W1123 23:23:55.032331 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.032357 kubelet[2655]: E1123 23:23:55.032344 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.032661 kubelet[2655]: E1123 23:23:55.032648 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.032661 kubelet[2655]: W1123 23:23:55.032661 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.032760 kubelet[2655]: E1123 23:23:55.032672 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.033047 kubelet[2655]: E1123 23:23:55.033032 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.033079 kubelet[2655]: W1123 23:23:55.033048 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.033079 kubelet[2655]: E1123 23:23:55.033059 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.033386 kubelet[2655]: E1123 23:23:55.033368 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.033513 kubelet[2655]: W1123 23:23:55.033495 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.033906 kubelet[2655]: E1123 23:23:55.033885 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.034076 kubelet[2655]: E1123 23:23:55.034063 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.034105 kubelet[2655]: W1123 23:23:55.034076 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.034105 kubelet[2655]: E1123 23:23:55.034087 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.034278 kubelet[2655]: E1123 23:23:55.034262 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.034278 kubelet[2655]: W1123 23:23:55.034275 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.034344 kubelet[2655]: E1123 23:23:55.034284 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.034649 kubelet[2655]: E1123 23:23:55.034635 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.034683 kubelet[2655]: W1123 23:23:55.034650 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.034683 kubelet[2655]: E1123 23:23:55.034661 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.035031 kubelet[2655]: E1123 23:23:55.035015 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.035031 kubelet[2655]: W1123 23:23:55.035030 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.035106 kubelet[2655]: E1123 23:23:55.035041 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.035544 kubelet[2655]: E1123 23:23:55.035524 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.035579 kubelet[2655]: W1123 23:23:55.035548 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.035717 kubelet[2655]: E1123 23:23:55.035562 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.036152 kubelet[2655]: E1123 23:23:55.036136 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.036152 kubelet[2655]: W1123 23:23:55.036153 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.036234 kubelet[2655]: E1123 23:23:55.036165 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.036432 kubelet[2655]: E1123 23:23:55.036416 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.036468 kubelet[2655]: W1123 23:23:55.036431 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.036468 kubelet[2655]: E1123 23:23:55.036442 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.037199 kubelet[2655]: E1123 23:23:55.037175 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.037199 kubelet[2655]: W1123 23:23:55.037191 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.037318 kubelet[2655]: E1123 23:23:55.037299 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.037686 kubelet[2655]: E1123 23:23:55.037670 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.037841 kubelet[2655]: W1123 23:23:55.037761 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.037841 kubelet[2655]: E1123 23:23:55.037781 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.042100 kubelet[2655]: E1123 23:23:55.041997 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.042100 kubelet[2655]: W1123 23:23:55.042017 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.042100 kubelet[2655]: E1123 23:23:55.042032 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.042701 kubelet[2655]: E1123 23:23:55.042565 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.042701 kubelet[2655]: W1123 23:23:55.042584 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.042701 kubelet[2655]: E1123 23:23:55.042611 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.043059 kubelet[2655]: E1123 23:23:55.043043 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.043191 kubelet[2655]: W1123 23:23:55.043118 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.043191 kubelet[2655]: E1123 23:23:55.043142 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.043436 kubelet[2655]: E1123 23:23:55.043410 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.043436 kubelet[2655]: W1123 23:23:55.043432 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.043584 kubelet[2655]: E1123 23:23:55.043564 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.043825 kubelet[2655]: E1123 23:23:55.043806 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.043825 kubelet[2655]: W1123 23:23:55.043818 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.043825 kubelet[2655]: E1123 23:23:55.043833 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.044120 kubelet[2655]: E1123 23:23:55.044103 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.044120 kubelet[2655]: W1123 23:23:55.044118 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.044213 kubelet[2655]: E1123 23:23:55.044152 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.044350 kubelet[2655]: E1123 23:23:55.044304 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.044350 kubelet[2655]: W1123 23:23:55.044317 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.044350 kubelet[2655]: E1123 23:23:55.044342 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.044685 kubelet[2655]: E1123 23:23:55.044563 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.044685 kubelet[2655]: W1123 23:23:55.044604 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.044685 kubelet[2655]: E1123 23:23:55.044638 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.044887 kubelet[2655]: E1123 23:23:55.044799 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.044887 kubelet[2655]: W1123 23:23:55.044808 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.044887 kubelet[2655]: E1123 23:23:55.044824 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.045463 kubelet[2655]: E1123 23:23:55.045313 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.045463 kubelet[2655]: W1123 23:23:55.045329 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.045463 kubelet[2655]: E1123 23:23:55.045351 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.045657 kubelet[2655]: E1123 23:23:55.045644 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.045847 kubelet[2655]: W1123 23:23:55.045744 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.045970 kubelet[2655]: E1123 23:23:55.045918 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.046059 kubelet[2655]: E1123 23:23:55.046041 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.046097 kubelet[2655]: W1123 23:23:55.046058 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.046097 kubelet[2655]: E1123 23:23:55.046077 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.047634 kubelet[2655]: E1123 23:23:55.047527 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.047634 kubelet[2655]: W1123 23:23:55.047542 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.047634 kubelet[2655]: E1123 23:23:55.047594 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.047750 kubelet[2655]: E1123 23:23:55.047737 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.047750 kubelet[2655]: W1123 23:23:55.047747 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.047945 kubelet[2655]: E1123 23:23:55.047811 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.047945 kubelet[2655]: E1123 23:23:55.047936 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.047945 kubelet[2655]: W1123 23:23:55.047945 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.048025 kubelet[2655]: E1123 23:23:55.048009 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.048151 kubelet[2655]: E1123 23:23:55.048096 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.048151 kubelet[2655]: W1123 23:23:55.048107 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.048151 kubelet[2655]: E1123 23:23:55.048118 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.048383 kubelet[2655]: E1123 23:23:55.048359 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.048383 kubelet[2655]: W1123 23:23:55.048374 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.048462 kubelet[2655]: E1123 23:23:55.048386 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.048629 kubelet[2655]: E1123 23:23:55.048610 2655 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 23 23:23:55.048629 kubelet[2655]: W1123 23:23:55.048629 2655 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 23 23:23:55.048698 kubelet[2655]: E1123 23:23:55.048642 2655 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 23 23:23:55.108159 containerd[1527]: time="2025-11-23T23:23:55.108099616Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:55.108634 containerd[1527]: time="2025-11-23T23:23:55.108609766Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=4266741" Nov 23 23:23:55.109418 containerd[1527]: time="2025-11-23T23:23:55.109372191Z" level=info msg="ImageCreate event name:\"sha256:90ff755393144dc5a3c05f95ffe1a3ecd2f89b98ecf36d9e4721471b80af4640\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:55.111519 containerd[1527]: time="2025-11-23T23:23:55.111331311Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:55.112127 containerd[1527]: time="2025-11-23T23:23:55.111994698Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:90ff755393144dc5a3c05f95ffe1a3ecd2f89b98ecf36d9e4721471b80af4640\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5636392\" in 1.18066494s" Nov 23 23:23:55.112127 containerd[1527]: time="2025-11-23T23:23:55.112024977Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:90ff755393144dc5a3c05f95ffe1a3ecd2f89b98ecf36d9e4721471b80af4640\"" Nov 23 23:23:55.118258 containerd[1527]: time="2025-11-23T23:23:55.118222371Z" level=info msg="CreateContainer within sandbox \"63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Nov 23 23:23:55.125611 containerd[1527]: time="2025-11-23T23:23:55.125140231Z" level=info msg="Container 2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:23:55.134074 containerd[1527]: time="2025-11-23T23:23:55.134023811Z" level=info msg="CreateContainer within sandbox \"63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e\"" Nov 23 23:23:55.134905 containerd[1527]: time="2025-11-23T23:23:55.134862474Z" level=info msg="StartContainer for \"2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e\"" Nov 23 23:23:55.136685 containerd[1527]: time="2025-11-23T23:23:55.136618279Z" level=info msg="connecting to shim 2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e" address="unix:///run/containerd/s/42d2b95fbb2b2e6d11cf4073a2f15e963ce0fdbf425861dfae92df00910752f4" protocol=ttrpc version=3 Nov 23 23:23:55.162656 systemd[1]: Started cri-containerd-2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e.scope - libcontainer container 2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e. Nov 23 23:23:55.253758 containerd[1527]: time="2025-11-23T23:23:55.253059600Z" level=info msg="StartContainer for \"2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e\" returns successfully" Nov 23 23:23:55.256517 systemd[1]: cri-containerd-2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e.scope: Deactivated successfully. Nov 23 23:23:55.277223 containerd[1527]: time="2025-11-23T23:23:55.276954316Z" level=info msg="received container exit event container_id:\"2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e\" id:\"2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e\" pid:3357 exited_at:{seconds:1763940235 nanos:271620584}" Nov 23 23:23:55.321228 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2fd5b47e32cd2a573171d05c8090cd8e08b5fbf3e58e8f5d23b780b871857b5e-rootfs.mount: Deactivated successfully. Nov 23 23:23:55.881976 kubelet[2655]: E1123 23:23:55.881927 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:23:55.989284 kubelet[2655]: I1123 23:23:55.989226 2655 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 23:23:55.990831 containerd[1527]: time="2025-11-23T23:23:55.990772056Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Nov 23 23:23:57.881526 kubelet[2655]: E1123 23:23:57.881465 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:23:59.203945 containerd[1527]: time="2025-11-23T23:23:59.203902378Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:59.204402 containerd[1527]: time="2025-11-23T23:23:59.204376970Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=65925816" Nov 23 23:23:59.205255 containerd[1527]: time="2025-11-23T23:23:59.205230715Z" level=info msg="ImageCreate event name:\"sha256:e60d442b6496497355efdf45eaa3ea72f5a2b28a5187aeab33442933f3c735d2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:59.207050 containerd[1527]: time="2025-11-23T23:23:59.207005045Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:23:59.208166 containerd[1527]: time="2025-11-23T23:23:59.207871950Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:e60d442b6496497355efdf45eaa3ea72f5a2b28a5187aeab33442933f3c735d2\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"67295507\" in 3.217035455s" Nov 23 23:23:59.208166 containerd[1527]: time="2025-11-23T23:23:59.207903989Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:e60d442b6496497355efdf45eaa3ea72f5a2b28a5187aeab33442933f3c735d2\"" Nov 23 23:23:59.211610 containerd[1527]: time="2025-11-23T23:23:59.211568287Z" level=info msg="CreateContainer within sandbox \"63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Nov 23 23:23:59.218839 containerd[1527]: time="2025-11-23T23:23:59.218502248Z" level=info msg="Container e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:23:59.225715 containerd[1527]: time="2025-11-23T23:23:59.225661046Z" level=info msg="CreateContainer within sandbox \"63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7\"" Nov 23 23:23:59.226191 containerd[1527]: time="2025-11-23T23:23:59.226144518Z" level=info msg="StartContainer for \"e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7\"" Nov 23 23:23:59.227946 containerd[1527]: time="2025-11-23T23:23:59.227907968Z" level=info msg="connecting to shim e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7" address="unix:///run/containerd/s/42d2b95fbb2b2e6d11cf4073a2f15e963ce0fdbf425861dfae92df00910752f4" protocol=ttrpc version=3 Nov 23 23:23:59.249678 systemd[1]: Started cri-containerd-e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7.scope - libcontainer container e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7. Nov 23 23:23:59.320319 containerd[1527]: time="2025-11-23T23:23:59.320277549Z" level=info msg="StartContainer for \"e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7\" returns successfully" Nov 23 23:23:59.859626 systemd[1]: cri-containerd-e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7.scope: Deactivated successfully. Nov 23 23:23:59.859893 systemd[1]: cri-containerd-e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7.scope: Consumed 431ms CPU time, 176.5M memory peak, 2.3M read from disk, 165.9M written to disk. Nov 23 23:23:59.864418 kubelet[2655]: I1123 23:23:59.864392 2655 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Nov 23 23:23:59.870719 containerd[1527]: time="2025-11-23T23:23:59.870682904Z" level=info msg="received container exit event container_id:\"e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7\" id:\"e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7\" pid:3418 exited_at:{seconds:1763940239 nanos:869106851}" Nov 23 23:23:59.890777 systemd[1]: Created slice kubepods-besteffort-pod82559865_90d3_4a8e_83a3_002c0c14990f.slice - libcontainer container kubepods-besteffort-pod82559865_90d3_4a8e_83a3_002c0c14990f.slice. Nov 23 23:23:59.899914 containerd[1527]: time="2025-11-23T23:23:59.899838486Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-b2nkd,Uid:82559865-90d3-4a8e-83a3-002c0c14990f,Namespace:calico-system,Attempt:0,}" Nov 23 23:23:59.909203 systemd[1]: Created slice kubepods-burstable-pod5b58e49f_cc99_463b_9194_304f2691d0a8.slice - libcontainer container kubepods-burstable-pod5b58e49f_cc99_463b_9194_304f2691d0a8.slice. Nov 23 23:23:59.922978 systemd[1]: Created slice kubepods-besteffort-podd63258be_88c5_4ec0_93e4_199929ede838.slice - libcontainer container kubepods-besteffort-podd63258be_88c5_4ec0_93e4_199929ede838.slice. Nov 23 23:23:59.935590 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e0ba244e4c2da352a9ae3dbea34eb2d852c13aa6889224a30064e5da7b8a1dc7-rootfs.mount: Deactivated successfully. Nov 23 23:23:59.942064 systemd[1]: Created slice kubepods-besteffort-pod927c8d49_e51e_462f_a42c_bb251141354f.slice - libcontainer container kubepods-besteffort-pod927c8d49_e51e_462f_a42c_bb251141354f.slice. Nov 23 23:23:59.948242 systemd[1]: Created slice kubepods-besteffort-pod856c0a9c_c169_4a6e_b144_99d3c9e7525a.slice - libcontainer container kubepods-besteffort-pod856c0a9c_c169_4a6e_b144_99d3c9e7525a.slice. Nov 23 23:23:59.953875 systemd[1]: Created slice kubepods-burstable-pod3dd78f68_bc34_49dc_9213_487d2a49c0ac.slice - libcontainer container kubepods-burstable-pod3dd78f68_bc34_49dc_9213_487d2a49c0ac.slice. Nov 23 23:23:59.965099 systemd[1]: Created slice kubepods-besteffort-pod5c544408_aa7d_4999_b85e_a233becf4f04.slice - libcontainer container kubepods-besteffort-pod5c544408_aa7d_4999_b85e_a233becf4f04.slice. Nov 23 23:23:59.969938 systemd[1]: Created slice kubepods-besteffort-pod17e162c7_08ce_4bc1_bbf1_49ed806bcede.slice - libcontainer container kubepods-besteffort-pod17e162c7_08ce_4bc1_bbf1_49ed806bcede.slice. Nov 23 23:23:59.974234 systemd[1]: Created slice kubepods-besteffort-pod31b181b6_1069_492f_a0d6_9078f4ef9b6a.slice - libcontainer container kubepods-besteffort-pod31b181b6_1069_492f_a0d6_9078f4ef9b6a.slice. Nov 23 23:23:59.976834 kubelet[2655]: I1123 23:23:59.976792 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/856c0a9c-c169-4a6e-b144-99d3c9e7525a-whisker-ca-bundle\") pod \"whisker-6566df9f4-r9ttx\" (UID: \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\") " pod="calico-system/whisker-6566df9f4-r9ttx" Nov 23 23:23:59.976834 kubelet[2655]: I1123 23:23:59.976834 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/5c544408-aa7d-4999-b85e-a233becf4f04-goldmane-key-pair\") pod \"goldmane-666569f655-4h52k\" (UID: \"5c544408-aa7d-4999-b85e-a233becf4f04\") " pod="calico-system/goldmane-666569f655-4h52k" Nov 23 23:23:59.976960 kubelet[2655]: I1123 23:23:59.976853 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whk88\" (UniqueName: \"kubernetes.io/projected/927c8d49-e51e-462f-a42c-bb251141354f-kube-api-access-whk88\") pod \"calico-apiserver-846996596c-s6g44\" (UID: \"927c8d49-e51e-462f-a42c-bb251141354f\") " pod="calico-apiserver/calico-apiserver-846996596c-s6g44" Nov 23 23:23:59.976960 kubelet[2655]: I1123 23:23:59.976876 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31b181b6-1069-492f-a0d6-9078f4ef9b6a-tigera-ca-bundle\") pod \"calico-kube-controllers-5c9f86c848-rx7m5\" (UID: \"31b181b6-1069-492f-a0d6-9078f4ef9b6a\") " pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" Nov 23 23:23:59.976960 kubelet[2655]: I1123 23:23:59.976895 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c544408-aa7d-4999-b85e-a233becf4f04-goldmane-ca-bundle\") pod \"goldmane-666569f655-4h52k\" (UID: \"5c544408-aa7d-4999-b85e-a233becf4f04\") " pod="calico-system/goldmane-666569f655-4h52k" Nov 23 23:23:59.976960 kubelet[2655]: I1123 23:23:59.976914 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr98w\" (UniqueName: \"kubernetes.io/projected/d63258be-88c5-4ec0-93e4-199929ede838-kube-api-access-vr98w\") pod \"calico-apiserver-59795bc4c9-sq9fp\" (UID: \"d63258be-88c5-4ec0-93e4-199929ede838\") " pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" Nov 23 23:23:59.976960 kubelet[2655]: I1123 23:23:59.976930 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/17e162c7-08ce-4bc1-bbf1-49ed806bcede-calico-apiserver-certs\") pod \"calico-apiserver-59795bc4c9-k44tv\" (UID: \"17e162c7-08ce-4bc1-bbf1-49ed806bcede\") " pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" Nov 23 23:23:59.977073 kubelet[2655]: I1123 23:23:59.976964 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b58e49f-cc99-463b-9194-304f2691d0a8-config-volume\") pod \"coredns-668d6bf9bc-clqmv\" (UID: \"5b58e49f-cc99-463b-9194-304f2691d0a8\") " pod="kube-system/coredns-668d6bf9bc-clqmv" Nov 23 23:23:59.977073 kubelet[2655]: I1123 23:23:59.976983 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3dd78f68-bc34-49dc-9213-487d2a49c0ac-config-volume\") pod \"coredns-668d6bf9bc-vwvns\" (UID: \"3dd78f68-bc34-49dc-9213-487d2a49c0ac\") " pod="kube-system/coredns-668d6bf9bc-vwvns" Nov 23 23:23:59.977073 kubelet[2655]: I1123 23:23:59.976999 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/d63258be-88c5-4ec0-93e4-199929ede838-calico-apiserver-certs\") pod \"calico-apiserver-59795bc4c9-sq9fp\" (UID: \"d63258be-88c5-4ec0-93e4-199929ede838\") " pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" Nov 23 23:23:59.977073 kubelet[2655]: I1123 23:23:59.977014 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw4gr\" (UniqueName: \"kubernetes.io/projected/5b58e49f-cc99-463b-9194-304f2691d0a8-kube-api-access-jw4gr\") pod \"coredns-668d6bf9bc-clqmv\" (UID: \"5b58e49f-cc99-463b-9194-304f2691d0a8\") " pod="kube-system/coredns-668d6bf9bc-clqmv" Nov 23 23:23:59.977073 kubelet[2655]: I1123 23:23:59.977028 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94mjb\" (UniqueName: \"kubernetes.io/projected/5c544408-aa7d-4999-b85e-a233becf4f04-kube-api-access-94mjb\") pod \"goldmane-666569f655-4h52k\" (UID: \"5c544408-aa7d-4999-b85e-a233becf4f04\") " pod="calico-system/goldmane-666569f655-4h52k" Nov 23 23:23:59.977171 kubelet[2655]: I1123 23:23:59.977048 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j26cb\" (UniqueName: \"kubernetes.io/projected/856c0a9c-c169-4a6e-b144-99d3c9e7525a-kube-api-access-j26cb\") pod \"whisker-6566df9f4-r9ttx\" (UID: \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\") " pod="calico-system/whisker-6566df9f4-r9ttx" Nov 23 23:23:59.977171 kubelet[2655]: I1123 23:23:59.977065 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddfs2\" (UniqueName: \"kubernetes.io/projected/31b181b6-1069-492f-a0d6-9078f4ef9b6a-kube-api-access-ddfs2\") pod \"calico-kube-controllers-5c9f86c848-rx7m5\" (UID: \"31b181b6-1069-492f-a0d6-9078f4ef9b6a\") " pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" Nov 23 23:23:59.979497 kubelet[2655]: I1123 23:23:59.979247 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/856c0a9c-c169-4a6e-b144-99d3c9e7525a-whisker-backend-key-pair\") pod \"whisker-6566df9f4-r9ttx\" (UID: \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\") " pod="calico-system/whisker-6566df9f4-r9ttx" Nov 23 23:23:59.979497 kubelet[2655]: I1123 23:23:59.979331 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/927c8d49-e51e-462f-a42c-bb251141354f-calico-apiserver-certs\") pod \"calico-apiserver-846996596c-s6g44\" (UID: \"927c8d49-e51e-462f-a42c-bb251141354f\") " pod="calico-apiserver/calico-apiserver-846996596c-s6g44" Nov 23 23:23:59.979497 kubelet[2655]: I1123 23:23:59.979354 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c544408-aa7d-4999-b85e-a233becf4f04-config\") pod \"goldmane-666569f655-4h52k\" (UID: \"5c544408-aa7d-4999-b85e-a233becf4f04\") " pod="calico-system/goldmane-666569f655-4h52k" Nov 23 23:23:59.979497 kubelet[2655]: I1123 23:23:59.979376 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6bf4\" (UniqueName: \"kubernetes.io/projected/3dd78f68-bc34-49dc-9213-487d2a49c0ac-kube-api-access-v6bf4\") pod \"coredns-668d6bf9bc-vwvns\" (UID: \"3dd78f68-bc34-49dc-9213-487d2a49c0ac\") " pod="kube-system/coredns-668d6bf9bc-vwvns" Nov 23 23:23:59.979497 kubelet[2655]: I1123 23:23:59.979397 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lckd7\" (UniqueName: \"kubernetes.io/projected/17e162c7-08ce-4bc1-bbf1-49ed806bcede-kube-api-access-lckd7\") pod \"calico-apiserver-59795bc4c9-k44tv\" (UID: \"17e162c7-08ce-4bc1-bbf1-49ed806bcede\") " pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" Nov 23 23:24:00.071987 containerd[1527]: time="2025-11-23T23:24:00.071917992Z" level=error msg="Failed to destroy network for sandbox \"325656d59ac8c6715d632796141f6ffa6cf67a78b901ab6695b22a8a7f53d123\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.073215 containerd[1527]: time="2025-11-23T23:24:00.073173052Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-b2nkd,Uid:82559865-90d3-4a8e-83a3-002c0c14990f,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"325656d59ac8c6715d632796141f6ffa6cf67a78b901ab6695b22a8a7f53d123\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.077420 kubelet[2655]: E1123 23:24:00.077267 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"325656d59ac8c6715d632796141f6ffa6cf67a78b901ab6695b22a8a7f53d123\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.077420 kubelet[2655]: E1123 23:24:00.077363 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"325656d59ac8c6715d632796141f6ffa6cf67a78b901ab6695b22a8a7f53d123\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-b2nkd" Nov 23 23:24:00.077420 kubelet[2655]: E1123 23:24:00.077382 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"325656d59ac8c6715d632796141f6ffa6cf67a78b901ab6695b22a8a7f53d123\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-b2nkd" Nov 23 23:24:00.077997 kubelet[2655]: E1123 23:24:00.077640 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-b2nkd_calico-system(82559865-90d3-4a8e-83a3-002c0c14990f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-b2nkd_calico-system(82559865-90d3-4a8e-83a3-002c0c14990f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"325656d59ac8c6715d632796141f6ffa6cf67a78b901ab6695b22a8a7f53d123\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:24:00.222015 containerd[1527]: time="2025-11-23T23:24:00.221120903Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-clqmv,Uid:5b58e49f-cc99-463b-9194-304f2691d0a8,Namespace:kube-system,Attempt:0,}" Nov 23 23:24:00.230922 systemd[1]: run-netns-cni\x2d23041c52\x2d4460\x2dc92b\x2de221\x2df6b3b1c150be.mount: Deactivated successfully. Nov 23 23:24:00.233981 containerd[1527]: time="2025-11-23T23:24:00.233811775Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59795bc4c9-sq9fp,Uid:d63258be-88c5-4ec0-93e4-199929ede838,Namespace:calico-apiserver,Attempt:0,}" Nov 23 23:24:00.257607 containerd[1527]: time="2025-11-23T23:24:00.257357028Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-846996596c-s6g44,Uid:927c8d49-e51e-462f-a42c-bb251141354f,Namespace:calico-apiserver,Attempt:0,}" Nov 23 23:24:00.262598 containerd[1527]: time="2025-11-23T23:24:00.262562623Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6566df9f4-r9ttx,Uid:856c0a9c-c169-4a6e-b144-99d3c9e7525a,Namespace:calico-system,Attempt:0,}" Nov 23 23:24:00.263588 containerd[1527]: time="2025-11-23T23:24:00.263201332Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vwvns,Uid:3dd78f68-bc34-49dc-9213-487d2a49c0ac,Namespace:kube-system,Attempt:0,}" Nov 23 23:24:00.268653 containerd[1527]: time="2025-11-23T23:24:00.268573204Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-4h52k,Uid:5c544408-aa7d-4999-b85e-a233becf4f04,Namespace:calico-system,Attempt:0,}" Nov 23 23:24:00.273404 containerd[1527]: time="2025-11-23T23:24:00.273372525Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59795bc4c9-k44tv,Uid:17e162c7-08ce-4bc1-bbf1-49ed806bcede,Namespace:calico-apiserver,Attempt:0,}" Nov 23 23:24:00.280291 containerd[1527]: time="2025-11-23T23:24:00.280230453Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5c9f86c848-rx7m5,Uid:31b181b6-1069-492f-a0d6-9078f4ef9b6a,Namespace:calico-system,Attempt:0,}" Nov 23 23:24:00.303206 containerd[1527]: time="2025-11-23T23:24:00.303156396Z" level=error msg="Failed to destroy network for sandbox \"e0da0e9215d6b358691fb91c378d2b0a311312e809e1395d46d7b991f4c44ed3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.307106 containerd[1527]: time="2025-11-23T23:24:00.307060732Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-clqmv,Uid:5b58e49f-cc99-463b-9194-304f2691d0a8,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e0da0e9215d6b358691fb91c378d2b0a311312e809e1395d46d7b991f4c44ed3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.307836 kubelet[2655]: E1123 23:24:00.307459 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e0da0e9215d6b358691fb91c378d2b0a311312e809e1395d46d7b991f4c44ed3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.307836 kubelet[2655]: E1123 23:24:00.307538 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e0da0e9215d6b358691fb91c378d2b0a311312e809e1395d46d7b991f4c44ed3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-clqmv" Nov 23 23:24:00.307836 kubelet[2655]: E1123 23:24:00.307557 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e0da0e9215d6b358691fb91c378d2b0a311312e809e1395d46d7b991f4c44ed3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-clqmv" Nov 23 23:24:00.307973 kubelet[2655]: E1123 23:24:00.307600 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-clqmv_kube-system(5b58e49f-cc99-463b-9194-304f2691d0a8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-clqmv_kube-system(5b58e49f-cc99-463b-9194-304f2691d0a8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e0da0e9215d6b358691fb91c378d2b0a311312e809e1395d46d7b991f4c44ed3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-clqmv" podUID="5b58e49f-cc99-463b-9194-304f2691d0a8" Nov 23 23:24:00.325402 containerd[1527]: time="2025-11-23T23:24:00.325340712Z" level=error msg="Failed to destroy network for sandbox \"3c313b1179d4cc5e5858174023b1592775bfca98d043260ead2a34905d024719\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.330153 containerd[1527]: time="2025-11-23T23:24:00.330088674Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59795bc4c9-sq9fp,Uid:d63258be-88c5-4ec0-93e4-199929ede838,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c313b1179d4cc5e5858174023b1592775bfca98d043260ead2a34905d024719\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.330702 kubelet[2655]: E1123 23:24:00.330372 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c313b1179d4cc5e5858174023b1592775bfca98d043260ead2a34905d024719\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.330702 kubelet[2655]: E1123 23:24:00.330431 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c313b1179d4cc5e5858174023b1592775bfca98d043260ead2a34905d024719\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" Nov 23 23:24:00.330702 kubelet[2655]: E1123 23:24:00.330449 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c313b1179d4cc5e5858174023b1592775bfca98d043260ead2a34905d024719\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" Nov 23 23:24:00.330809 kubelet[2655]: E1123 23:24:00.330568 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-59795bc4c9-sq9fp_calico-apiserver(d63258be-88c5-4ec0-93e4-199929ede838)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-59795bc4c9-sq9fp_calico-apiserver(d63258be-88c5-4ec0-93e4-199929ede838)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3c313b1179d4cc5e5858174023b1592775bfca98d043260ead2a34905d024719\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" podUID="d63258be-88c5-4ec0-93e4-199929ede838" Nov 23 23:24:00.346698 containerd[1527]: time="2025-11-23T23:24:00.346552204Z" level=error msg="Failed to destroy network for sandbox \"03057b0f93b313ef776614fa4045ad3eb4407aa1468e27cea4bd919dfde12dbb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.349578 containerd[1527]: time="2025-11-23T23:24:00.349523195Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-846996596c-s6g44,Uid:927c8d49-e51e-462f-a42c-bb251141354f,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"03057b0f93b313ef776614fa4045ad3eb4407aa1468e27cea4bd919dfde12dbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.349801 kubelet[2655]: E1123 23:24:00.349757 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"03057b0f93b313ef776614fa4045ad3eb4407aa1468e27cea4bd919dfde12dbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.349859 kubelet[2655]: E1123 23:24:00.349822 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"03057b0f93b313ef776614fa4045ad3eb4407aa1468e27cea4bd919dfde12dbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-846996596c-s6g44" Nov 23 23:24:00.349859 kubelet[2655]: E1123 23:24:00.349845 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"03057b0f93b313ef776614fa4045ad3eb4407aa1468e27cea4bd919dfde12dbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-846996596c-s6g44" Nov 23 23:24:00.349903 kubelet[2655]: E1123 23:24:00.349884 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-846996596c-s6g44_calico-apiserver(927c8d49-e51e-462f-a42c-bb251141354f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-846996596c-s6g44_calico-apiserver(927c8d49-e51e-462f-a42c-bb251141354f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"03057b0f93b313ef776614fa4045ad3eb4407aa1468e27cea4bd919dfde12dbb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-846996596c-s6g44" podUID="927c8d49-e51e-462f-a42c-bb251141354f" Nov 23 23:24:00.370173 containerd[1527]: time="2025-11-23T23:24:00.369960300Z" level=error msg="Failed to destroy network for sandbox \"f83c567a50364c713aa456d60988a14f1e54c49e5513a3d0a261691f9ba820bd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.372766 containerd[1527]: time="2025-11-23T23:24:00.372720015Z" level=error msg="Failed to destroy network for sandbox \"cc98d737e1535abdc032421f5fe70f854d229e686be8ecb084ac28e8fd120e63\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.373177 containerd[1527]: time="2025-11-23T23:24:00.373134088Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59795bc4c9-k44tv,Uid:17e162c7-08ce-4bc1-bbf1-49ed806bcede,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f83c567a50364c713aa456d60988a14f1e54c49e5513a3d0a261691f9ba820bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.373585 kubelet[2655]: E1123 23:24:00.373474 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f83c567a50364c713aa456d60988a14f1e54c49e5513a3d0a261691f9ba820bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.373707 kubelet[2655]: E1123 23:24:00.373633 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f83c567a50364c713aa456d60988a14f1e54c49e5513a3d0a261691f9ba820bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" Nov 23 23:24:00.373707 kubelet[2655]: E1123 23:24:00.373682 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f83c567a50364c713aa456d60988a14f1e54c49e5513a3d0a261691f9ba820bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" Nov 23 23:24:00.374081 kubelet[2655]: E1123 23:24:00.373861 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-59795bc4c9-k44tv_calico-apiserver(17e162c7-08ce-4bc1-bbf1-49ed806bcede)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-59795bc4c9-k44tv_calico-apiserver(17e162c7-08ce-4bc1-bbf1-49ed806bcede)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f83c567a50364c713aa456d60988a14f1e54c49e5513a3d0a261691f9ba820bd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" podUID="17e162c7-08ce-4bc1-bbf1-49ed806bcede" Nov 23 23:24:00.374150 containerd[1527]: time="2025-11-23T23:24:00.373906435Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vwvns,Uid:3dd78f68-bc34-49dc-9213-487d2a49c0ac,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc98d737e1535abdc032421f5fe70f854d229e686be8ecb084ac28e8fd120e63\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.374652 kubelet[2655]: E1123 23:24:00.374301 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc98d737e1535abdc032421f5fe70f854d229e686be8ecb084ac28e8fd120e63\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.374652 kubelet[2655]: E1123 23:24:00.374555 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc98d737e1535abdc032421f5fe70f854d229e686be8ecb084ac28e8fd120e63\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vwvns" Nov 23 23:24:00.374652 kubelet[2655]: E1123 23:24:00.374572 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc98d737e1535abdc032421f5fe70f854d229e686be8ecb084ac28e8fd120e63\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vwvns" Nov 23 23:24:00.374778 kubelet[2655]: E1123 23:24:00.374613 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-vwvns_kube-system(3dd78f68-bc34-49dc-9213-487d2a49c0ac)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-vwvns_kube-system(3dd78f68-bc34-49dc-9213-487d2a49c0ac)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cc98d737e1535abdc032421f5fe70f854d229e686be8ecb084ac28e8fd120e63\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-vwvns" podUID="3dd78f68-bc34-49dc-9213-487d2a49c0ac" Nov 23 23:24:00.376842 containerd[1527]: time="2025-11-23T23:24:00.376805467Z" level=error msg="Failed to destroy network for sandbox \"bde22d0e5f6b4a4d4de2435ab2725764609d722fe922cb78dfc0207111431046\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.379552 containerd[1527]: time="2025-11-23T23:24:00.379391465Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6566df9f4-r9ttx,Uid:856c0a9c-c169-4a6e-b144-99d3c9e7525a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bde22d0e5f6b4a4d4de2435ab2725764609d722fe922cb78dfc0207111431046\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.380059 kubelet[2655]: E1123 23:24:00.379980 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bde22d0e5f6b4a4d4de2435ab2725764609d722fe922cb78dfc0207111431046\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.380169 kubelet[2655]: E1123 23:24:00.380028 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bde22d0e5f6b4a4d4de2435ab2725764609d722fe922cb78dfc0207111431046\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6566df9f4-r9ttx" Nov 23 23:24:00.380169 kubelet[2655]: E1123 23:24:00.380138 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bde22d0e5f6b4a4d4de2435ab2725764609d722fe922cb78dfc0207111431046\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6566df9f4-r9ttx" Nov 23 23:24:00.380333 kubelet[2655]: E1123 23:24:00.380309 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-6566df9f4-r9ttx_calico-system(856c0a9c-c169-4a6e-b144-99d3c9e7525a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-6566df9f4-r9ttx_calico-system(856c0a9c-c169-4a6e-b144-99d3c9e7525a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bde22d0e5f6b4a4d4de2435ab2725764609d722fe922cb78dfc0207111431046\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-6566df9f4-r9ttx" podUID="856c0a9c-c169-4a6e-b144-99d3c9e7525a" Nov 23 23:24:00.384229 containerd[1527]: time="2025-11-23T23:24:00.384189786Z" level=error msg="Failed to destroy network for sandbox \"8d4aeeff58221940948a48548631be61c733763a62f4baaba76c266a188648ca\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.385463 containerd[1527]: time="2025-11-23T23:24:00.385099451Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-4h52k,Uid:5c544408-aa7d-4999-b85e-a233becf4f04,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8d4aeeff58221940948a48548631be61c733763a62f4baaba76c266a188648ca\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.385463 containerd[1527]: time="2025-11-23T23:24:00.385239249Z" level=error msg="Failed to destroy network for sandbox \"509d1235f8ba171431d78cd122ac36614f9510f53be9aecb7d09f2e6ec94d69d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.385584 kubelet[2655]: E1123 23:24:00.385254 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8d4aeeff58221940948a48548631be61c733763a62f4baaba76c266a188648ca\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.385584 kubelet[2655]: E1123 23:24:00.385308 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8d4aeeff58221940948a48548631be61c733763a62f4baaba76c266a188648ca\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-4h52k" Nov 23 23:24:00.385584 kubelet[2655]: E1123 23:24:00.385390 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8d4aeeff58221940948a48548631be61c733763a62f4baaba76c266a188648ca\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-4h52k" Nov 23 23:24:00.385665 kubelet[2655]: E1123 23:24:00.385435 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-4h52k_calico-system(5c544408-aa7d-4999-b85e-a233becf4f04)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-4h52k_calico-system(5c544408-aa7d-4999-b85e-a233becf4f04)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8d4aeeff58221940948a48548631be61c733763a62f4baaba76c266a188648ca\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-4h52k" podUID="5c544408-aa7d-4999-b85e-a233becf4f04" Nov 23 23:24:00.386178 containerd[1527]: time="2025-11-23T23:24:00.386140274Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5c9f86c848-rx7m5,Uid:31b181b6-1069-492f-a0d6-9078f4ef9b6a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"509d1235f8ba171431d78cd122ac36614f9510f53be9aecb7d09f2e6ec94d69d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.386389 kubelet[2655]: E1123 23:24:00.386354 2655 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"509d1235f8ba171431d78cd122ac36614f9510f53be9aecb7d09f2e6ec94d69d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 23 23:24:00.386695 kubelet[2655]: E1123 23:24:00.386673 2655 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"509d1235f8ba171431d78cd122ac36614f9510f53be9aecb7d09f2e6ec94d69d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" Nov 23 23:24:00.386857 kubelet[2655]: E1123 23:24:00.386759 2655 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"509d1235f8ba171431d78cd122ac36614f9510f53be9aecb7d09f2e6ec94d69d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" Nov 23 23:24:00.386929 kubelet[2655]: E1123 23:24:00.386807 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5c9f86c848-rx7m5_calico-system(31b181b6-1069-492f-a0d6-9078f4ef9b6a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5c9f86c848-rx7m5_calico-system(31b181b6-1069-492f-a0d6-9078f4ef9b6a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"509d1235f8ba171431d78cd122ac36614f9510f53be9aecb7d09f2e6ec94d69d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" podUID="31b181b6-1069-492f-a0d6-9078f4ef9b6a" Nov 23 23:24:01.006110 containerd[1527]: time="2025-11-23T23:24:01.006050981Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Nov 23 23:24:01.220198 systemd[1]: run-netns-cni\x2d27377659\x2da42b\x2dcd89\x2dd237\x2dab87d2cedc00.mount: Deactivated successfully. Nov 23 23:24:01.220296 systemd[1]: run-netns-cni\x2d6d6ab0ab\x2d4e88\x2dc563\x2d461d\x2d6544652742e8.mount: Deactivated successfully. Nov 23 23:24:01.220340 systemd[1]: run-netns-cni\x2d3c46d93b\x2d4814\x2dec55\x2d7192\x2de93d811ace62.mount: Deactivated successfully. Nov 23 23:24:01.220384 systemd[1]: run-netns-cni\x2d30453601\x2d2992\x2d09b0\x2de1c7\x2de0f3b4e158f1.mount: Deactivated successfully. Nov 23 23:24:04.747230 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1557300373.mount: Deactivated successfully. Nov 23 23:24:04.988143 containerd[1527]: time="2025-11-23T23:24:04.988086981Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=150934562" Nov 23 23:24:04.991943 containerd[1527]: time="2025-11-23T23:24:04.991893367Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:43a5290057a103af76996c108856f92ed902f34573d7a864f55f15b8aaf4683b\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"150934424\" in 3.985779907s" Nov 23 23:24:04.991943 containerd[1527]: time="2025-11-23T23:24:04.991935406Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:43a5290057a103af76996c108856f92ed902f34573d7a864f55f15b8aaf4683b\"" Nov 23 23:24:04.999956 containerd[1527]: time="2025-11-23T23:24:04.999802375Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:24:05.001034 containerd[1527]: time="2025-11-23T23:24:05.000996639Z" level=info msg="ImageCreate event name:\"sha256:43a5290057a103af76996c108856f92ed902f34573d7a864f55f15b8aaf4683b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:24:05.002298 containerd[1527]: time="2025-11-23T23:24:05.002215902Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 23 23:24:05.004659 containerd[1527]: time="2025-11-23T23:24:05.004626909Z" level=info msg="CreateContainer within sandbox \"63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Nov 23 23:24:05.019511 containerd[1527]: time="2025-11-23T23:24:05.018995353Z" level=info msg="Container 9c57e005fc2871b69b54c6054b81be7c87db6b665e3f50922aed90c36f29f435: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:24:05.030041 containerd[1527]: time="2025-11-23T23:24:05.029986363Z" level=info msg="CreateContainer within sandbox \"63391bd85c7254f295c54697ce1396ee0fd4548e01133ef72ad2cba2c6831638\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"9c57e005fc2871b69b54c6054b81be7c87db6b665e3f50922aed90c36f29f435\"" Nov 23 23:24:05.030733 containerd[1527]: time="2025-11-23T23:24:05.030708474Z" level=info msg="StartContainer for \"9c57e005fc2871b69b54c6054b81be7c87db6b665e3f50922aed90c36f29f435\"" Nov 23 23:24:05.032164 containerd[1527]: time="2025-11-23T23:24:05.032137534Z" level=info msg="connecting to shim 9c57e005fc2871b69b54c6054b81be7c87db6b665e3f50922aed90c36f29f435" address="unix:///run/containerd/s/42d2b95fbb2b2e6d11cf4073a2f15e963ce0fdbf425861dfae92df00910752f4" protocol=ttrpc version=3 Nov 23 23:24:05.057675 systemd[1]: Started cri-containerd-9c57e005fc2871b69b54c6054b81be7c87db6b665e3f50922aed90c36f29f435.scope - libcontainer container 9c57e005fc2871b69b54c6054b81be7c87db6b665e3f50922aed90c36f29f435. Nov 23 23:24:05.138956 containerd[1527]: time="2025-11-23T23:24:05.138846399Z" level=info msg="StartContainer for \"9c57e005fc2871b69b54c6054b81be7c87db6b665e3f50922aed90c36f29f435\" returns successfully" Nov 23 23:24:05.259450 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Nov 23 23:24:05.259601 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Nov 23 23:24:05.412298 kubelet[2655]: I1123 23:24:05.412228 2655 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j26cb\" (UniqueName: \"kubernetes.io/projected/856c0a9c-c169-4a6e-b144-99d3c9e7525a-kube-api-access-j26cb\") pod \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\" (UID: \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\") " Nov 23 23:24:05.412894 kubelet[2655]: I1123 23:24:05.412756 2655 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/856c0a9c-c169-4a6e-b144-99d3c9e7525a-whisker-ca-bundle\") pod \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\" (UID: \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\") " Nov 23 23:24:05.412894 kubelet[2655]: I1123 23:24:05.412800 2655 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/856c0a9c-c169-4a6e-b144-99d3c9e7525a-whisker-backend-key-pair\") pod \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\" (UID: \"856c0a9c-c169-4a6e-b144-99d3c9e7525a\") " Nov 23 23:24:05.424138 kubelet[2655]: I1123 23:24:05.424085 2655 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/856c0a9c-c169-4a6e-b144-99d3c9e7525a-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "856c0a9c-c169-4a6e-b144-99d3c9e7525a" (UID: "856c0a9c-c169-4a6e-b144-99d3c9e7525a"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Nov 23 23:24:05.425632 kubelet[2655]: I1123 23:24:05.425596 2655 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/856c0a9c-c169-4a6e-b144-99d3c9e7525a-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "856c0a9c-c169-4a6e-b144-99d3c9e7525a" (UID: "856c0a9c-c169-4a6e-b144-99d3c9e7525a"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Nov 23 23:24:05.426104 kubelet[2655]: I1123 23:24:05.426061 2655 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/856c0a9c-c169-4a6e-b144-99d3c9e7525a-kube-api-access-j26cb" (OuterVolumeSpecName: "kube-api-access-j26cb") pod "856c0a9c-c169-4a6e-b144-99d3c9e7525a" (UID: "856c0a9c-c169-4a6e-b144-99d3c9e7525a"). InnerVolumeSpecName "kube-api-access-j26cb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Nov 23 23:24:05.514288 kubelet[2655]: I1123 23:24:05.514228 2655 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/856c0a9c-c169-4a6e-b144-99d3c9e7525a-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Nov 23 23:24:05.514288 kubelet[2655]: I1123 23:24:05.514271 2655 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/856c0a9c-c169-4a6e-b144-99d3c9e7525a-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Nov 23 23:24:05.514288 kubelet[2655]: I1123 23:24:05.514281 2655 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-j26cb\" (UniqueName: \"kubernetes.io/projected/856c0a9c-c169-4a6e-b144-99d3c9e7525a-kube-api-access-j26cb\") on node \"localhost\" DevicePath \"\"" Nov 23 23:24:05.747015 systemd[1]: var-lib-kubelet-pods-856c0a9c\x2dc169\x2d4a6e\x2db144\x2d99d3c9e7525a-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dj26cb.mount: Deactivated successfully. Nov 23 23:24:05.747115 systemd[1]: var-lib-kubelet-pods-856c0a9c\x2dc169\x2d4a6e\x2db144\x2d99d3c9e7525a-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Nov 23 23:24:06.024741 systemd[1]: Removed slice kubepods-besteffort-pod856c0a9c_c169_4a6e_b144_99d3c9e7525a.slice - libcontainer container kubepods-besteffort-pod856c0a9c_c169_4a6e_b144_99d3c9e7525a.slice. Nov 23 23:24:06.037207 kubelet[2655]: I1123 23:24:06.037131 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-g2qw5" podStartSLOduration=2.138828257 podStartE2EDuration="15.037112089s" podCreationTimestamp="2025-11-23 23:23:51 +0000 UTC" firstStartedPulling="2025-11-23 23:23:52.095191033 +0000 UTC m=+21.285089458" lastFinishedPulling="2025-11-23 23:24:04.993474865 +0000 UTC m=+34.183373290" observedRunningTime="2025-11-23 23:24:06.03705113 +0000 UTC m=+35.226949555" watchObservedRunningTime="2025-11-23 23:24:06.037112089 +0000 UTC m=+35.227010474" Nov 23 23:24:06.087631 systemd[1]: Created slice kubepods-besteffort-pod47273dbf_2dbb_4b82_b6aa_30c6db81392e.slice - libcontainer container kubepods-besteffort-pod47273dbf_2dbb_4b82_b6aa_30c6db81392e.slice. Nov 23 23:24:06.119865 kubelet[2655]: I1123 23:24:06.119828 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgd29\" (UniqueName: \"kubernetes.io/projected/47273dbf-2dbb-4b82-b6aa-30c6db81392e-kube-api-access-kgd29\") pod \"whisker-5b9c899964-zqqgk\" (UID: \"47273dbf-2dbb-4b82-b6aa-30c6db81392e\") " pod="calico-system/whisker-5b9c899964-zqqgk" Nov 23 23:24:06.120004 kubelet[2655]: I1123 23:24:06.119880 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47273dbf-2dbb-4b82-b6aa-30c6db81392e-whisker-ca-bundle\") pod \"whisker-5b9c899964-zqqgk\" (UID: \"47273dbf-2dbb-4b82-b6aa-30c6db81392e\") " pod="calico-system/whisker-5b9c899964-zqqgk" Nov 23 23:24:06.120004 kubelet[2655]: I1123 23:24:06.119904 2655 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/47273dbf-2dbb-4b82-b6aa-30c6db81392e-whisker-backend-key-pair\") pod \"whisker-5b9c899964-zqqgk\" (UID: \"47273dbf-2dbb-4b82-b6aa-30c6db81392e\") " pod="calico-system/whisker-5b9c899964-zqqgk" Nov 23 23:24:06.392572 containerd[1527]: time="2025-11-23T23:24:06.392515406Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5b9c899964-zqqgk,Uid:47273dbf-2dbb-4b82-b6aa-30c6db81392e,Namespace:calico-system,Attempt:0,}" Nov 23 23:24:06.587835 systemd-networkd[1446]: calie25d867b57d: Link UP Nov 23 23:24:06.588283 systemd-networkd[1446]: calie25d867b57d: Gained carrier Nov 23 23:24:06.611658 containerd[1527]: 2025-11-23 23:24:06.427 [INFO][3829] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Nov 23 23:24:06.611658 containerd[1527]: 2025-11-23 23:24:06.466 [INFO][3829] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--5b9c899964--zqqgk-eth0 whisker-5b9c899964- calico-system 47273dbf-2dbb-4b82-b6aa-30c6db81392e 873 0 2025-11-23 23:24:06 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:5b9c899964 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-5b9c899964-zqqgk eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calie25d867b57d [] [] }} ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Namespace="calico-system" Pod="whisker-5b9c899964-zqqgk" WorkloadEndpoint="localhost-k8s-whisker--5b9c899964--zqqgk-" Nov 23 23:24:06.611658 containerd[1527]: 2025-11-23 23:24:06.466 [INFO][3829] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Namespace="calico-system" Pod="whisker-5b9c899964-zqqgk" WorkloadEndpoint="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" Nov 23 23:24:06.611658 containerd[1527]: 2025-11-23 23:24:06.527 [INFO][3844] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" HandleID="k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Workload="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.528 [INFO][3844] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" HandleID="k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Workload="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000322b30), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-5b9c899964-zqqgk", "timestamp":"2025-11-23 23:24:06.527874342 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.528 [INFO][3844] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.528 [INFO][3844] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.528 [INFO][3844] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.538 [INFO][3844] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" host="localhost" Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.542 [INFO][3844] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.547 [INFO][3844] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.548 [INFO][3844] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.550 [INFO][3844] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:06.612118 containerd[1527]: 2025-11-23 23:24:06.550 [INFO][3844] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" host="localhost" Nov 23 23:24:06.612333 containerd[1527]: 2025-11-23 23:24:06.552 [INFO][3844] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12 Nov 23 23:24:06.612333 containerd[1527]: 2025-11-23 23:24:06.556 [INFO][3844] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" host="localhost" Nov 23 23:24:06.612333 containerd[1527]: 2025-11-23 23:24:06.567 [INFO][3844] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" host="localhost" Nov 23 23:24:06.612333 containerd[1527]: 2025-11-23 23:24:06.567 [INFO][3844] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" host="localhost" Nov 23 23:24:06.612333 containerd[1527]: 2025-11-23 23:24:06.568 [INFO][3844] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:06.612333 containerd[1527]: 2025-11-23 23:24:06.568 [INFO][3844] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" HandleID="k8s-pod-network.3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Workload="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" Nov 23 23:24:06.612447 containerd[1527]: 2025-11-23 23:24:06.572 [INFO][3829] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Namespace="calico-system" Pod="whisker-5b9c899964-zqqgk" WorkloadEndpoint="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--5b9c899964--zqqgk-eth0", GenerateName:"whisker-5b9c899964-", Namespace:"calico-system", SelfLink:"", UID:"47273dbf-2dbb-4b82-b6aa-30c6db81392e", ResourceVersion:"873", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 24, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5b9c899964", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-5b9c899964-zqqgk", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calie25d867b57d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:06.612447 containerd[1527]: 2025-11-23 23:24:06.572 [INFO][3829] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Namespace="calico-system" Pod="whisker-5b9c899964-zqqgk" WorkloadEndpoint="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" Nov 23 23:24:06.612579 containerd[1527]: 2025-11-23 23:24:06.572 [INFO][3829] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie25d867b57d ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Namespace="calico-system" Pod="whisker-5b9c899964-zqqgk" WorkloadEndpoint="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" Nov 23 23:24:06.612579 containerd[1527]: 2025-11-23 23:24:06.588 [INFO][3829] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Namespace="calico-system" Pod="whisker-5b9c899964-zqqgk" WorkloadEndpoint="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" Nov 23 23:24:06.613327 containerd[1527]: 2025-11-23 23:24:06.589 [INFO][3829] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Namespace="calico-system" Pod="whisker-5b9c899964-zqqgk" WorkloadEndpoint="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--5b9c899964--zqqgk-eth0", GenerateName:"whisker-5b9c899964-", Namespace:"calico-system", SelfLink:"", UID:"47273dbf-2dbb-4b82-b6aa-30c6db81392e", ResourceVersion:"873", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 24, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5b9c899964", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12", Pod:"whisker-5b9c899964-zqqgk", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calie25d867b57d", MAC:"d6:50:7f:0c:d6:90", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:06.613430 containerd[1527]: 2025-11-23 23:24:06.600 [INFO][3829] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" Namespace="calico-system" Pod="whisker-5b9c899964-zqqgk" WorkloadEndpoint="localhost-k8s-whisker--5b9c899964--zqqgk-eth0" Nov 23 23:24:06.706182 containerd[1527]: time="2025-11-23T23:24:06.706043674Z" level=info msg="connecting to shim 3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12" address="unix:///run/containerd/s/b8002a195b6adeb40ea96d0c76d301ef753b780a5441d0602dc2f8bbc8dc2b20" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:06.740648 systemd[1]: Started cri-containerd-3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12.scope - libcontainer container 3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12. Nov 23 23:24:06.758862 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:06.786178 containerd[1527]: time="2025-11-23T23:24:06.786123899Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5b9c899964-zqqgk,Uid:47273dbf-2dbb-4b82-b6aa-30c6db81392e,Namespace:calico-system,Attempt:0,} returns sandbox id \"3877143dfd756a0d35363959f4342b38782733bb2d52647476dcf1318e337c12\"" Nov 23 23:24:06.787682 containerd[1527]: time="2025-11-23T23:24:06.787579040Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 23 23:24:06.883854 kubelet[2655]: I1123 23:24:06.883680 2655 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="856c0a9c-c169-4a6e-b144-99d3c9e7525a" path="/var/lib/kubelet/pods/856c0a9c-c169-4a6e-b144-99d3c9e7525a/volumes" Nov 23 23:24:06.991706 containerd[1527]: time="2025-11-23T23:24:06.991363795Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:07.003935 containerd[1527]: time="2025-11-23T23:24:07.003877111Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 23 23:24:07.004005 containerd[1527]: time="2025-11-23T23:24:07.003945910Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 23 23:24:07.004224 kubelet[2655]: E1123 23:24:07.004161 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 23 23:24:07.004287 kubelet[2655]: E1123 23:24:07.004233 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 23 23:24:07.005844 kubelet[2655]: E1123 23:24:07.005741 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:605c7f6a192d4d0899f32eea354f7105,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kgd29,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5b9c899964-zqqgk_calico-system(47273dbf-2dbb-4b82-b6aa-30c6db81392e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:07.008192 containerd[1527]: time="2025-11-23T23:24:07.008164496Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 23 23:24:07.229794 containerd[1527]: time="2025-11-23T23:24:07.229734151Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:07.258752 containerd[1527]: time="2025-11-23T23:24:07.258397946Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 23 23:24:07.258752 containerd[1527]: time="2025-11-23T23:24:07.258442945Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 23 23:24:07.258871 kubelet[2655]: E1123 23:24:07.258721 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 23 23:24:07.258871 kubelet[2655]: E1123 23:24:07.258771 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 23 23:24:07.258934 kubelet[2655]: E1123 23:24:07.258888 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kgd29,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5b9c899964-zqqgk_calico-system(47273dbf-2dbb-4b82-b6aa-30c6db81392e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:07.260147 kubelet[2655]: E1123 23:24:07.260075 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5b9c899964-zqqgk" podUID="47273dbf-2dbb-4b82-b6aa-30c6db81392e" Nov 23 23:24:07.843707 systemd-networkd[1446]: calie25d867b57d: Gained IPv6LL Nov 23 23:24:08.028898 kubelet[2655]: E1123 23:24:08.028853 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5b9c899964-zqqgk" podUID="47273dbf-2dbb-4b82-b6aa-30c6db81392e" Nov 23 23:24:08.873820 kubelet[2655]: I1123 23:24:08.873780 2655 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 23:24:10.135046 systemd-networkd[1446]: vxlan.calico: Link UP Nov 23 23:24:10.135057 systemd-networkd[1446]: vxlan.calico: Gained carrier Nov 23 23:24:10.884069 containerd[1527]: time="2025-11-23T23:24:10.883737702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59795bc4c9-sq9fp,Uid:d63258be-88c5-4ec0-93e4-199929ede838,Namespace:calico-apiserver,Attempt:0,}" Nov 23 23:24:11.009408 systemd-networkd[1446]: calie6338fcc659: Link UP Nov 23 23:24:11.009926 systemd-networkd[1446]: calie6338fcc659: Gained carrier Nov 23 23:24:11.056657 containerd[1527]: 2025-11-23 23:24:10.922 [INFO][4233] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0 calico-apiserver-59795bc4c9- calico-apiserver d63258be-88c5-4ec0-93e4-199929ede838 808 0 2025-11-23 23:23:45 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:59795bc4c9 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-59795bc4c9-sq9fp eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calie6338fcc659 [] [] }} ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-sq9fp" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-" Nov 23 23:24:11.056657 containerd[1527]: 2025-11-23 23:24:10.922 [INFO][4233] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-sq9fp" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" Nov 23 23:24:11.056657 containerd[1527]: 2025-11-23 23:24:10.948 [INFO][4248] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" HandleID="k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Workload="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.948 [INFO][4248] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" HandleID="k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Workload="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004d690), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-59795bc4c9-sq9fp", "timestamp":"2025-11-23 23:24:10.948274593 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.948 [INFO][4248] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.948 [INFO][4248] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.948 [INFO][4248] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.959 [INFO][4248] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" host="localhost" Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.965 [INFO][4248] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.975 [INFO][4248] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.978 [INFO][4248] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.981 [INFO][4248] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:11.057186 containerd[1527]: 2025-11-23 23:24:10.981 [INFO][4248] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" host="localhost" Nov 23 23:24:11.057389 containerd[1527]: 2025-11-23 23:24:10.982 [INFO][4248] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea Nov 23 23:24:11.057389 containerd[1527]: 2025-11-23 23:24:10.990 [INFO][4248] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" host="localhost" Nov 23 23:24:11.057389 containerd[1527]: 2025-11-23 23:24:11.000 [INFO][4248] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" host="localhost" Nov 23 23:24:11.057389 containerd[1527]: 2025-11-23 23:24:11.000 [INFO][4248] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" host="localhost" Nov 23 23:24:11.057389 containerd[1527]: 2025-11-23 23:24:11.000 [INFO][4248] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:11.057389 containerd[1527]: 2025-11-23 23:24:11.000 [INFO][4248] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" HandleID="k8s-pod-network.d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Workload="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" Nov 23 23:24:11.057511 containerd[1527]: 2025-11-23 23:24:11.005 [INFO][4233] cni-plugin/k8s.go 418: Populated endpoint ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-sq9fp" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0", GenerateName:"calico-apiserver-59795bc4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"d63258be-88c5-4ec0-93e4-199929ede838", ResourceVersion:"808", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59795bc4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-59795bc4c9-sq9fp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie6338fcc659", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:11.057570 containerd[1527]: 2025-11-23 23:24:11.005 [INFO][4233] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-sq9fp" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" Nov 23 23:24:11.057570 containerd[1527]: 2025-11-23 23:24:11.005 [INFO][4233] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie6338fcc659 ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-sq9fp" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" Nov 23 23:24:11.057570 containerd[1527]: 2025-11-23 23:24:11.010 [INFO][4233] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-sq9fp" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" Nov 23 23:24:11.058347 containerd[1527]: 2025-11-23 23:24:11.011 [INFO][4233] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-sq9fp" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0", GenerateName:"calico-apiserver-59795bc4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"d63258be-88c5-4ec0-93e4-199929ede838", ResourceVersion:"808", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59795bc4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea", Pod:"calico-apiserver-59795bc4c9-sq9fp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie6338fcc659", MAC:"a2:c8:90:a5:99:b2", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:11.058630 containerd[1527]: 2025-11-23 23:24:11.052 [INFO][4233] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-sq9fp" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--sq9fp-eth0" Nov 23 23:24:11.115220 containerd[1527]: time="2025-11-23T23:24:11.115179411Z" level=info msg="connecting to shim d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea" address="unix:///run/containerd/s/7ca1101f7e80d8d18cb83c8f83d8a21486897847ba66d65afdf012ac9770a02b" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:11.149401 systemd[1]: Started cri-containerd-d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea.scope - libcontainer container d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea. Nov 23 23:24:11.170201 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:11.195310 containerd[1527]: time="2025-11-23T23:24:11.195245067Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59795bc4c9-sq9fp,Uid:d63258be-88c5-4ec0-93e4-199929ede838,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"d9fcf7e375a8a27db9d5061f1e5adef22c1b3e2de6ed1ec8267a253298d8c8ea\"" Nov 23 23:24:11.196786 containerd[1527]: time="2025-11-23T23:24:11.196762130Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 23 23:24:11.398666 containerd[1527]: time="2025-11-23T23:24:11.398613852Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:11.399576 containerd[1527]: time="2025-11-23T23:24:11.399545201Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 23 23:24:11.399736 containerd[1527]: time="2025-11-23T23:24:11.399618280Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 23 23:24:11.399779 kubelet[2655]: E1123 23:24:11.399733 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:11.400013 kubelet[2655]: E1123 23:24:11.399777 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:11.400942 kubelet[2655]: E1123 23:24:11.400885 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vr98w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59795bc4c9-sq9fp_calico-apiserver(d63258be-88c5-4ec0-93e4-199929ede838): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:11.402098 kubelet[2655]: E1123 23:24:11.402062 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" podUID="d63258be-88c5-4ec0-93e4-199929ede838" Nov 23 23:24:11.505083 systemd[1]: Started sshd@7-10.0.0.134:22-10.0.0.1:53214.service - OpenSSH per-connection server daemon (10.0.0.1:53214). Nov 23 23:24:11.581160 sshd[4313]: Accepted publickey for core from 10.0.0.1 port 53214 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:11.582422 sshd-session[4313]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:11.586687 systemd-logind[1513]: New session 8 of user core. Nov 23 23:24:11.594639 systemd[1]: Started session-8.scope - Session 8 of User core. Nov 23 23:24:11.740602 sshd[4316]: Connection closed by 10.0.0.1 port 53214 Nov 23 23:24:11.741075 sshd-session[4313]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:11.745293 systemd[1]: sshd@7-10.0.0.134:22-10.0.0.1:53214.service: Deactivated successfully. Nov 23 23:24:11.746901 systemd[1]: session-8.scope: Deactivated successfully. Nov 23 23:24:11.747758 systemd-logind[1513]: Session 8 logged out. Waiting for processes to exit. Nov 23 23:24:11.749143 systemd-logind[1513]: Removed session 8. Nov 23 23:24:11.882988 containerd[1527]: time="2025-11-23T23:24:11.882936664Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-4h52k,Uid:5c544408-aa7d-4999-b85e-a233becf4f04,Namespace:calico-system,Attempt:0,}" Nov 23 23:24:12.043276 systemd-networkd[1446]: calic0005e302a9: Link UP Nov 23 23:24:12.044017 systemd-networkd[1446]: calic0005e302a9: Gained carrier Nov 23 23:24:12.050900 kubelet[2655]: E1123 23:24:12.050859 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" podUID="d63258be-88c5-4ec0-93e4-199929ede838" Nov 23 23:24:12.060954 containerd[1527]: 2025-11-23 23:24:11.962 [INFO][4329] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--666569f655--4h52k-eth0 goldmane-666569f655- calico-system 5c544408-aa7d-4999-b85e-a233becf4f04 809 0 2025-11-23 23:23:49 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-666569f655-4h52k eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calic0005e302a9 [] [] }} ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Namespace="calico-system" Pod="goldmane-666569f655-4h52k" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--4h52k-" Nov 23 23:24:12.060954 containerd[1527]: 2025-11-23 23:24:11.963 [INFO][4329] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Namespace="calico-system" Pod="goldmane-666569f655-4h52k" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--4h52k-eth0" Nov 23 23:24:12.060954 containerd[1527]: 2025-11-23 23:24:11.990 [INFO][4346] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" HandleID="k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Workload="localhost-k8s-goldmane--666569f655--4h52k-eth0" Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:11.990 [INFO][4346] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" HandleID="k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Workload="localhost-k8s-goldmane--666569f655--4h52k-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002d3680), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-666569f655-4h52k", "timestamp":"2025-11-23 23:24:11.990174294 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:11.990 [INFO][4346] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:11.990 [INFO][4346] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:11.990 [INFO][4346] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:12.000 [INFO][4346] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" host="localhost" Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:12.005 [INFO][4346] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:12.012 [INFO][4346] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:12.014 [INFO][4346] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:12.017 [INFO][4346] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:12.061327 containerd[1527]: 2025-11-23 23:24:12.017 [INFO][4346] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" host="localhost" Nov 23 23:24:12.061624 containerd[1527]: 2025-11-23 23:24:12.018 [INFO][4346] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130 Nov 23 23:24:12.061624 containerd[1527]: 2025-11-23 23:24:12.030 [INFO][4346] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" host="localhost" Nov 23 23:24:12.061624 containerd[1527]: 2025-11-23 23:24:12.036 [INFO][4346] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" host="localhost" Nov 23 23:24:12.061624 containerd[1527]: 2025-11-23 23:24:12.036 [INFO][4346] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" host="localhost" Nov 23 23:24:12.061624 containerd[1527]: 2025-11-23 23:24:12.036 [INFO][4346] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:12.061624 containerd[1527]: 2025-11-23 23:24:12.036 [INFO][4346] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" HandleID="k8s-pod-network.f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Workload="localhost-k8s-goldmane--666569f655--4h52k-eth0" Nov 23 23:24:12.061993 containerd[1527]: 2025-11-23 23:24:12.040 [INFO][4329] cni-plugin/k8s.go 418: Populated endpoint ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Namespace="calico-system" Pod="goldmane-666569f655-4h52k" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--4h52k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--4h52k-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"5c544408-aa7d-4999-b85e-a233becf4f04", ResourceVersion:"809", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-666569f655-4h52k", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calic0005e302a9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:12.061993 containerd[1527]: 2025-11-23 23:24:12.040 [INFO][4329] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Namespace="calico-system" Pod="goldmane-666569f655-4h52k" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--4h52k-eth0" Nov 23 23:24:12.062085 containerd[1527]: 2025-11-23 23:24:12.040 [INFO][4329] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic0005e302a9 ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Namespace="calico-system" Pod="goldmane-666569f655-4h52k" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--4h52k-eth0" Nov 23 23:24:12.062085 containerd[1527]: 2025-11-23 23:24:12.044 [INFO][4329] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Namespace="calico-system" Pod="goldmane-666569f655-4h52k" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--4h52k-eth0" Nov 23 23:24:12.062129 containerd[1527]: 2025-11-23 23:24:12.044 [INFO][4329] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Namespace="calico-system" Pod="goldmane-666569f655-4h52k" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--4h52k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--4h52k-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"5c544408-aa7d-4999-b85e-a233becf4f04", ResourceVersion:"809", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130", Pod:"goldmane-666569f655-4h52k", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calic0005e302a9", MAC:"76:1b:98:20:38:a5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:12.062176 containerd[1527]: 2025-11-23 23:24:12.057 [INFO][4329] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" Namespace="calico-system" Pod="goldmane-666569f655-4h52k" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--4h52k-eth0" Nov 23 23:24:12.068602 systemd-networkd[1446]: vxlan.calico: Gained IPv6LL Nov 23 23:24:12.088362 containerd[1527]: time="2025-11-23T23:24:12.087993337Z" level=info msg="connecting to shim f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130" address="unix:///run/containerd/s/5697efb5e8a462eddbe98c7726be95a162a2d262ccb9ee405fb777ecb3a8baf2" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:12.112656 systemd[1]: Started cri-containerd-f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130.scope - libcontainer container f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130. Nov 23 23:24:12.123895 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:12.145700 containerd[1527]: time="2025-11-23T23:24:12.145368867Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-4h52k,Uid:5c544408-aa7d-4999-b85e-a233becf4f04,Namespace:calico-system,Attempt:0,} returns sandbox id \"f54ec8cae97cbab21c92240f4eff396e1fa1664b05bae7321c0d0cb3b0224130\"" Nov 23 23:24:12.147120 containerd[1527]: time="2025-11-23T23:24:12.147091168Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 23 23:24:12.195637 systemd-networkd[1446]: calie6338fcc659: Gained IPv6LL Nov 23 23:24:12.352474 containerd[1527]: time="2025-11-23T23:24:12.352427114Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:12.353381 containerd[1527]: time="2025-11-23T23:24:12.353326304Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 23 23:24:12.353426 containerd[1527]: time="2025-11-23T23:24:12.353366063Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 23 23:24:12.353623 kubelet[2655]: E1123 23:24:12.353565 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 23 23:24:12.353672 kubelet[2655]: E1123 23:24:12.353637 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 23 23:24:12.353812 kubelet[2655]: E1123 23:24:12.353770 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-94mjb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-4h52k_calico-system(5c544408-aa7d-4999-b85e-a233becf4f04): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:12.356001 kubelet[2655]: E1123 23:24:12.355950 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-4h52k" podUID="5c544408-aa7d-4999-b85e-a233becf4f04" Nov 23 23:24:12.882329 containerd[1527]: time="2025-11-23T23:24:12.882272577Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5c9f86c848-rx7m5,Uid:31b181b6-1069-492f-a0d6-9078f4ef9b6a,Namespace:calico-system,Attempt:0,}" Nov 23 23:24:12.994969 systemd-networkd[1446]: cali291f061fb4e: Link UP Nov 23 23:24:12.996956 systemd-networkd[1446]: cali291f061fb4e: Gained carrier Nov 23 23:24:13.011171 containerd[1527]: 2025-11-23 23:24:12.922 [INFO][4415] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0 calico-kube-controllers-5c9f86c848- calico-system 31b181b6-1069-492f-a0d6-9078f4ef9b6a 802 0 2025-11-23 23:23:51 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5c9f86c848 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-5c9f86c848-rx7m5 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali291f061fb4e [] [] }} ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Namespace="calico-system" Pod="calico-kube-controllers-5c9f86c848-rx7m5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-" Nov 23 23:24:13.011171 containerd[1527]: 2025-11-23 23:24:12.922 [INFO][4415] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Namespace="calico-system" Pod="calico-kube-controllers-5c9f86c848-rx7m5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" Nov 23 23:24:13.011171 containerd[1527]: 2025-11-23 23:24:12.952 [INFO][4429] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" HandleID="k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Workload="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.952 [INFO][4429] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" HandleID="k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Workload="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004cbf0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-5c9f86c848-rx7m5", "timestamp":"2025-11-23 23:24:12.952036611 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.952 [INFO][4429] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.952 [INFO][4429] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.952 [INFO][4429] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.963 [INFO][4429] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" host="localhost" Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.968 [INFO][4429] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.972 [INFO][4429] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.974 [INFO][4429] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.976 [INFO][4429] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:13.011404 containerd[1527]: 2025-11-23 23:24:12.976 [INFO][4429] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" host="localhost" Nov 23 23:24:13.011936 containerd[1527]: 2025-11-23 23:24:12.977 [INFO][4429] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c Nov 23 23:24:13.011936 containerd[1527]: 2025-11-23 23:24:12.981 [INFO][4429] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" host="localhost" Nov 23 23:24:13.011936 containerd[1527]: 2025-11-23 23:24:12.987 [INFO][4429] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" host="localhost" Nov 23 23:24:13.011936 containerd[1527]: 2025-11-23 23:24:12.987 [INFO][4429] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" host="localhost" Nov 23 23:24:13.011936 containerd[1527]: 2025-11-23 23:24:12.987 [INFO][4429] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:13.011936 containerd[1527]: 2025-11-23 23:24:12.987 [INFO][4429] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" HandleID="k8s-pod-network.6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Workload="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" Nov 23 23:24:13.012096 containerd[1527]: 2025-11-23 23:24:12.991 [INFO][4415] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Namespace="calico-system" Pod="calico-kube-controllers-5c9f86c848-rx7m5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0", GenerateName:"calico-kube-controllers-5c9f86c848-", Namespace:"calico-system", SelfLink:"", UID:"31b181b6-1069-492f-a0d6-9078f4ef9b6a", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5c9f86c848", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-5c9f86c848-rx7m5", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali291f061fb4e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:13.012168 containerd[1527]: 2025-11-23 23:24:12.991 [INFO][4415] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Namespace="calico-system" Pod="calico-kube-controllers-5c9f86c848-rx7m5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" Nov 23 23:24:13.012168 containerd[1527]: 2025-11-23 23:24:12.991 [INFO][4415] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali291f061fb4e ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Namespace="calico-system" Pod="calico-kube-controllers-5c9f86c848-rx7m5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" Nov 23 23:24:13.012168 containerd[1527]: 2025-11-23 23:24:12.995 [INFO][4415] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Namespace="calico-system" Pod="calico-kube-controllers-5c9f86c848-rx7m5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" Nov 23 23:24:13.012314 containerd[1527]: 2025-11-23 23:24:12.996 [INFO][4415] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Namespace="calico-system" Pod="calico-kube-controllers-5c9f86c848-rx7m5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0", GenerateName:"calico-kube-controllers-5c9f86c848-", Namespace:"calico-system", SelfLink:"", UID:"31b181b6-1069-492f-a0d6-9078f4ef9b6a", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5c9f86c848", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c", Pod:"calico-kube-controllers-5c9f86c848-rx7m5", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali291f061fb4e", MAC:"0a:8f:93:7a:10:40", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:13.012391 containerd[1527]: 2025-11-23 23:24:13.005 [INFO][4415] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" Namespace="calico-system" Pod="calico-kube-controllers-5c9f86c848-rx7m5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5c9f86c848--rx7m5-eth0" Nov 23 23:24:13.043708 containerd[1527]: time="2025-11-23T23:24:13.043663777Z" level=info msg="connecting to shim 6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c" address="unix:///run/containerd/s/d0605946efeadfa58d4cca717a9cef4bde1d1515152b6bb388aef774ab36d31d" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:13.055671 kubelet[2655]: E1123 23:24:13.055631 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" podUID="d63258be-88c5-4ec0-93e4-199929ede838" Nov 23 23:24:13.071222 kubelet[2655]: E1123 23:24:13.071187 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-4h52k" podUID="5c544408-aa7d-4999-b85e-a233becf4f04" Nov 23 23:24:13.095652 systemd[1]: Started cri-containerd-6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c.scope - libcontainer container 6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c. Nov 23 23:24:13.109850 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:13.140579 containerd[1527]: time="2025-11-23T23:24:13.139919588Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5c9f86c848-rx7m5,Uid:31b181b6-1069-492f-a0d6-9078f4ef9b6a,Namespace:calico-system,Attempt:0,} returns sandbox id \"6303f0a46a35898cca25ebe7932c8d472d4f0b9603f74fced2f718548cf9d91c\"" Nov 23 23:24:13.142475 containerd[1527]: time="2025-11-23T23:24:13.142448841Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 23 23:24:13.349735 containerd[1527]: time="2025-11-23T23:24:13.349672067Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:13.350723 containerd[1527]: time="2025-11-23T23:24:13.350686456Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 23 23:24:13.350850 containerd[1527]: time="2025-11-23T23:24:13.350760535Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 23 23:24:13.350981 kubelet[2655]: E1123 23:24:13.350939 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 23 23:24:13.351054 kubelet[2655]: E1123 23:24:13.350992 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 23 23:24:13.351189 kubelet[2655]: E1123 23:24:13.351137 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ddfs2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5c9f86c848-rx7m5_calico-system(31b181b6-1069-492f-a0d6-9078f4ef9b6a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:13.352265 kubelet[2655]: E1123 23:24:13.352228 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" podUID="31b181b6-1069-492f-a0d6-9078f4ef9b6a" Nov 23 23:24:13.859705 systemd-networkd[1446]: calic0005e302a9: Gained IPv6LL Nov 23 23:24:14.058245 kubelet[2655]: E1123 23:24:14.058190 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-4h52k" podUID="5c544408-aa7d-4999-b85e-a233becf4f04" Nov 23 23:24:14.058623 kubelet[2655]: E1123 23:24:14.058549 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" podUID="31b181b6-1069-492f-a0d6-9078f4ef9b6a" Nov 23 23:24:14.115793 systemd-networkd[1446]: cali291f061fb4e: Gained IPv6LL Nov 23 23:24:14.882050 containerd[1527]: time="2025-11-23T23:24:14.881980729Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vwvns,Uid:3dd78f68-bc34-49dc-9213-487d2a49c0ac,Namespace:kube-system,Attempt:0,}" Nov 23 23:24:14.882609 containerd[1527]: time="2025-11-23T23:24:14.882435204Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-846996596c-s6g44,Uid:927c8d49-e51e-462f-a42c-bb251141354f,Namespace:calico-apiserver,Attempt:0,}" Nov 23 23:24:14.882609 containerd[1527]: time="2025-11-23T23:24:14.882547603Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-b2nkd,Uid:82559865-90d3-4a8e-83a3-002c0c14990f,Namespace:calico-system,Attempt:0,}" Nov 23 23:24:15.023328 systemd-networkd[1446]: calic23cb12fb4b: Link UP Nov 23 23:24:15.023456 systemd-networkd[1446]: calic23cb12fb4b: Gained carrier Nov 23 23:24:15.044555 containerd[1527]: 2025-11-23 23:24:14.944 [INFO][4504] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--b2nkd-eth0 csi-node-driver- calico-system 82559865-90d3-4a8e-83a3-002c0c14990f 703 0 2025-11-23 23:23:51 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-b2nkd eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calic23cb12fb4b [] [] }} ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Namespace="calico-system" Pod="csi-node-driver-b2nkd" WorkloadEndpoint="localhost-k8s-csi--node--driver--b2nkd-" Nov 23 23:24:15.044555 containerd[1527]: 2025-11-23 23:24:14.944 [INFO][4504] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Namespace="calico-system" Pod="csi-node-driver-b2nkd" WorkloadEndpoint="localhost-k8s-csi--node--driver--b2nkd-eth0" Nov 23 23:24:15.044555 containerd[1527]: 2025-11-23 23:24:14.975 [INFO][4545] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" HandleID="k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Workload="localhost-k8s-csi--node--driver--b2nkd-eth0" Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:14.975 [INFO][4545] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" HandleID="k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Workload="localhost-k8s-csi--node--driver--b2nkd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002c32c0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-b2nkd", "timestamp":"2025-11-23 23:24:14.975743952 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:14.976 [INFO][4545] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:14.976 [INFO][4545] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:14.976 [INFO][4545] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:14.988 [INFO][4545] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" host="localhost" Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:14.995 [INFO][4545] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:15.000 [INFO][4545] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:15.002 [INFO][4545] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:15.005 [INFO][4545] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:15.044798 containerd[1527]: 2025-11-23 23:24:15.005 [INFO][4545] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" host="localhost" Nov 23 23:24:15.044995 containerd[1527]: 2025-11-23 23:24:15.007 [INFO][4545] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5 Nov 23 23:24:15.044995 containerd[1527]: 2025-11-23 23:24:15.011 [INFO][4545] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" host="localhost" Nov 23 23:24:15.044995 containerd[1527]: 2025-11-23 23:24:15.018 [INFO][4545] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" host="localhost" Nov 23 23:24:15.044995 containerd[1527]: 2025-11-23 23:24:15.018 [INFO][4545] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" host="localhost" Nov 23 23:24:15.044995 containerd[1527]: 2025-11-23 23:24:15.018 [INFO][4545] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:15.044995 containerd[1527]: 2025-11-23 23:24:15.018 [INFO][4545] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" HandleID="k8s-pod-network.d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Workload="localhost-k8s-csi--node--driver--b2nkd-eth0" Nov 23 23:24:15.045124 containerd[1527]: 2025-11-23 23:24:15.020 [INFO][4504] cni-plugin/k8s.go 418: Populated endpoint ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Namespace="calico-system" Pod="csi-node-driver-b2nkd" WorkloadEndpoint="localhost-k8s-csi--node--driver--b2nkd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--b2nkd-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"82559865-90d3-4a8e-83a3-002c0c14990f", ResourceVersion:"703", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-b2nkd", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic23cb12fb4b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:15.045178 containerd[1527]: 2025-11-23 23:24:15.020 [INFO][4504] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Namespace="calico-system" Pod="csi-node-driver-b2nkd" WorkloadEndpoint="localhost-k8s-csi--node--driver--b2nkd-eth0" Nov 23 23:24:15.045178 containerd[1527]: 2025-11-23 23:24:15.020 [INFO][4504] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic23cb12fb4b ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Namespace="calico-system" Pod="csi-node-driver-b2nkd" WorkloadEndpoint="localhost-k8s-csi--node--driver--b2nkd-eth0" Nov 23 23:24:15.045178 containerd[1527]: 2025-11-23 23:24:15.022 [INFO][4504] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Namespace="calico-system" Pod="csi-node-driver-b2nkd" WorkloadEndpoint="localhost-k8s-csi--node--driver--b2nkd-eth0" Nov 23 23:24:15.045239 containerd[1527]: 2025-11-23 23:24:15.022 [INFO][4504] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Namespace="calico-system" Pod="csi-node-driver-b2nkd" WorkloadEndpoint="localhost-k8s-csi--node--driver--b2nkd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--b2nkd-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"82559865-90d3-4a8e-83a3-002c0c14990f", ResourceVersion:"703", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5", Pod:"csi-node-driver-b2nkd", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic23cb12fb4b", MAC:"9e:30:b0:45:63:2d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:15.045289 containerd[1527]: 2025-11-23 23:24:15.040 [INFO][4504] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" Namespace="calico-system" Pod="csi-node-driver-b2nkd" WorkloadEndpoint="localhost-k8s-csi--node--driver--b2nkd-eth0" Nov 23 23:24:15.061746 kubelet[2655]: E1123 23:24:15.061559 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" podUID="31b181b6-1069-492f-a0d6-9078f4ef9b6a" Nov 23 23:24:15.084256 containerd[1527]: time="2025-11-23T23:24:15.084160324Z" level=info msg="connecting to shim d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5" address="unix:///run/containerd/s/a59f5cacc679074307c0b4c227a45111127bc3d76e63edf934b7ff8724aa9641" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:15.110816 systemd[1]: Started cri-containerd-d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5.scope - libcontainer container d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5. Nov 23 23:24:15.128440 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:15.143636 systemd-networkd[1446]: cali01286db5d83: Link UP Nov 23 23:24:15.146160 systemd-networkd[1446]: cali01286db5d83: Gained carrier Nov 23 23:24:15.157951 containerd[1527]: time="2025-11-23T23:24:15.157902095Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-b2nkd,Uid:82559865-90d3-4a8e-83a3-002c0c14990f,Namespace:calico-system,Attempt:0,} returns sandbox id \"d7d2584fa312724333a310ce1ad4be98a746c1b9c25ceb345bab4acbf41d06b5\"" Nov 23 23:24:15.161016 containerd[1527]: time="2025-11-23T23:24:15.160710027Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 23 23:24:15.172888 containerd[1527]: 2025-11-23 23:24:14.946 [INFO][4502] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--846996596c--s6g44-eth0 calico-apiserver-846996596c- calico-apiserver 927c8d49-e51e-462f-a42c-bb251141354f 805 0 2025-11-23 23:23:47 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:846996596c projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-846996596c-s6g44 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali01286db5d83 [] [] }} ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Namespace="calico-apiserver" Pod="calico-apiserver-846996596c-s6g44" WorkloadEndpoint="localhost-k8s-calico--apiserver--846996596c--s6g44-" Nov 23 23:24:15.172888 containerd[1527]: 2025-11-23 23:24:14.946 [INFO][4502] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Namespace="calico-apiserver" Pod="calico-apiserver-846996596c-s6g44" WorkloadEndpoint="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" Nov 23 23:24:15.172888 containerd[1527]: 2025-11-23 23:24:14.980 [INFO][4547] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" HandleID="k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Workload="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:14.980 [INFO][4547] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" HandleID="k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Workload="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40001ab190), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-846996596c-s6g44", "timestamp":"2025-11-23 23:24:14.980042348 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:14.980 [INFO][4547] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:15.018 [INFO][4547] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:15.018 [INFO][4547] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:15.090 [INFO][4547] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" host="localhost" Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:15.097 [INFO][4547] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:15.106 [INFO][4547] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:15.109 [INFO][4547] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:15.112 [INFO][4547] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:15.173112 containerd[1527]: 2025-11-23 23:24:15.112 [INFO][4547] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" host="localhost" Nov 23 23:24:15.173380 containerd[1527]: 2025-11-23 23:24:15.115 [INFO][4547] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb Nov 23 23:24:15.173380 containerd[1527]: 2025-11-23 23:24:15.122 [INFO][4547] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" host="localhost" Nov 23 23:24:15.173380 containerd[1527]: 2025-11-23 23:24:15.132 [INFO][4547] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" host="localhost" Nov 23 23:24:15.173380 containerd[1527]: 2025-11-23 23:24:15.132 [INFO][4547] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" host="localhost" Nov 23 23:24:15.173380 containerd[1527]: 2025-11-23 23:24:15.132 [INFO][4547] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:15.173380 containerd[1527]: 2025-11-23 23:24:15.133 [INFO][4547] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" HandleID="k8s-pod-network.b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Workload="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" Nov 23 23:24:15.173505 containerd[1527]: 2025-11-23 23:24:15.139 [INFO][4502] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Namespace="calico-apiserver" Pod="calico-apiserver-846996596c-s6g44" WorkloadEndpoint="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--846996596c--s6g44-eth0", GenerateName:"calico-apiserver-846996596c-", Namespace:"calico-apiserver", SelfLink:"", UID:"927c8d49-e51e-462f-a42c-bb251141354f", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"846996596c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-846996596c-s6g44", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali01286db5d83", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:15.173555 containerd[1527]: 2025-11-23 23:24:15.139 [INFO][4502] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Namespace="calico-apiserver" Pod="calico-apiserver-846996596c-s6g44" WorkloadEndpoint="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" Nov 23 23:24:15.173555 containerd[1527]: 2025-11-23 23:24:15.139 [INFO][4502] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali01286db5d83 ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Namespace="calico-apiserver" Pod="calico-apiserver-846996596c-s6g44" WorkloadEndpoint="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" Nov 23 23:24:15.173555 containerd[1527]: 2025-11-23 23:24:15.147 [INFO][4502] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Namespace="calico-apiserver" Pod="calico-apiserver-846996596c-s6g44" WorkloadEndpoint="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" Nov 23 23:24:15.173623 containerd[1527]: 2025-11-23 23:24:15.149 [INFO][4502] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Namespace="calico-apiserver" Pod="calico-apiserver-846996596c-s6g44" WorkloadEndpoint="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--846996596c--s6g44-eth0", GenerateName:"calico-apiserver-846996596c-", Namespace:"calico-apiserver", SelfLink:"", UID:"927c8d49-e51e-462f-a42c-bb251141354f", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"846996596c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb", Pod:"calico-apiserver-846996596c-s6g44", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali01286db5d83", MAC:"3e:f5:fc:6d:23:8e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:15.173671 containerd[1527]: 2025-11-23 23:24:15.162 [INFO][4502] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" Namespace="calico-apiserver" Pod="calico-apiserver-846996596c-s6g44" WorkloadEndpoint="localhost-k8s-calico--apiserver--846996596c--s6g44-eth0" Nov 23 23:24:15.197949 containerd[1527]: time="2025-11-23T23:24:15.197904769Z" level=info msg="connecting to shim b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb" address="unix:///run/containerd/s/5e7c28ab0ab5e706c90326138485161480971b5f8e309606b97aba21f12973ea" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:15.228682 systemd[1]: Started cri-containerd-b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb.scope - libcontainer container b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb. Nov 23 23:24:15.238435 systemd-networkd[1446]: cali5a72d75a239: Link UP Nov 23 23:24:15.238871 systemd-networkd[1446]: cali5a72d75a239: Gained carrier Nov 23 23:24:15.256018 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:15.258536 containerd[1527]: 2025-11-23 23:24:14.952 [INFO][4496] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--vwvns-eth0 coredns-668d6bf9bc- kube-system 3dd78f68-bc34-49dc-9213-487d2a49c0ac 811 0 2025-11-23 23:23:36 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-vwvns eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali5a72d75a239 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Namespace="kube-system" Pod="coredns-668d6bf9bc-vwvns" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--vwvns-" Nov 23 23:24:15.258536 containerd[1527]: 2025-11-23 23:24:14.952 [INFO][4496] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Namespace="kube-system" Pod="coredns-668d6bf9bc-vwvns" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" Nov 23 23:24:15.258536 containerd[1527]: 2025-11-23 23:24:14.995 [INFO][4557] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" HandleID="k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Workload="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:14.995 [INFO][4557] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" HandleID="k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Workload="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003d20c0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-vwvns", "timestamp":"2025-11-23 23:24:14.995287909 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:14.995 [INFO][4557] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:15.132 [INFO][4557] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:15.132 [INFO][4557] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:15.192 [INFO][4557] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" host="localhost" Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:15.199 [INFO][4557] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:15.209 [INFO][4557] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:15.213 [INFO][4557] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:15.217 [INFO][4557] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:15.258736 containerd[1527]: 2025-11-23 23:24:15.217 [INFO][4557] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" host="localhost" Nov 23 23:24:15.258933 containerd[1527]: 2025-11-23 23:24:15.219 [INFO][4557] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d Nov 23 23:24:15.258933 containerd[1527]: 2025-11-23 23:24:15.225 [INFO][4557] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" host="localhost" Nov 23 23:24:15.258933 containerd[1527]: 2025-11-23 23:24:15.233 [INFO][4557] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" host="localhost" Nov 23 23:24:15.258933 containerd[1527]: 2025-11-23 23:24:15.233 [INFO][4557] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" host="localhost" Nov 23 23:24:15.258933 containerd[1527]: 2025-11-23 23:24:15.233 [INFO][4557] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:15.258933 containerd[1527]: 2025-11-23 23:24:15.233 [INFO][4557] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" HandleID="k8s-pod-network.db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Workload="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" Nov 23 23:24:15.259057 containerd[1527]: 2025-11-23 23:24:15.236 [INFO][4496] cni-plugin/k8s.go 418: Populated endpoint ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Namespace="kube-system" Pod="coredns-668d6bf9bc-vwvns" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--vwvns-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"3dd78f68-bc34-49dc-9213-487d2a49c0ac", ResourceVersion:"811", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-vwvns", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali5a72d75a239", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:15.259114 containerd[1527]: 2025-11-23 23:24:15.236 [INFO][4496] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Namespace="kube-system" Pod="coredns-668d6bf9bc-vwvns" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" Nov 23 23:24:15.259114 containerd[1527]: 2025-11-23 23:24:15.236 [INFO][4496] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5a72d75a239 ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Namespace="kube-system" Pod="coredns-668d6bf9bc-vwvns" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" Nov 23 23:24:15.259114 containerd[1527]: 2025-11-23 23:24:15.238 [INFO][4496] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Namespace="kube-system" Pod="coredns-668d6bf9bc-vwvns" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" Nov 23 23:24:15.259171 containerd[1527]: 2025-11-23 23:24:15.239 [INFO][4496] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Namespace="kube-system" Pod="coredns-668d6bf9bc-vwvns" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--vwvns-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"3dd78f68-bc34-49dc-9213-487d2a49c0ac", ResourceVersion:"811", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d", Pod:"coredns-668d6bf9bc-vwvns", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali5a72d75a239", MAC:"56:b9:9c:6f:47:0a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:15.259171 containerd[1527]: 2025-11-23 23:24:15.254 [INFO][4496] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" Namespace="kube-system" Pod="coredns-668d6bf9bc-vwvns" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--vwvns-eth0" Nov 23 23:24:15.292214 containerd[1527]: time="2025-11-23T23:24:15.292170131Z" level=info msg="connecting to shim db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d" address="unix:///run/containerd/s/d2287fcb7a8840dd391174a409ffdb0ea1e6822abcf11c8fc47957ec535129bf" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:15.300586 containerd[1527]: time="2025-11-23T23:24:15.300547646Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-846996596c-s6g44,Uid:927c8d49-e51e-462f-a42c-bb251141354f,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"b1ce3a3fd1dec6494cc8459e38f5f926077bcadabad93e3af88ba766b64632fb\"" Nov 23 23:24:15.328713 systemd[1]: Started cri-containerd-db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d.scope - libcontainer container db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d. Nov 23 23:24:15.347079 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:15.368461 containerd[1527]: time="2025-11-23T23:24:15.368419956Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vwvns,Uid:3dd78f68-bc34-49dc-9213-487d2a49c0ac,Namespace:kube-system,Attempt:0,} returns sandbox id \"db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d\"" Nov 23 23:24:15.371666 containerd[1527]: time="2025-11-23T23:24:15.371626164Z" level=info msg="CreateContainer within sandbox \"db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 23 23:24:15.377220 containerd[1527]: time="2025-11-23T23:24:15.377179107Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:15.379345 containerd[1527]: time="2025-11-23T23:24:15.379198767Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 23 23:24:15.379345 containerd[1527]: time="2025-11-23T23:24:15.379239726Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 23 23:24:15.379464 kubelet[2655]: E1123 23:24:15.379428 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 23 23:24:15.379550 kubelet[2655]: E1123 23:24:15.379477 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 23 23:24:15.379813 kubelet[2655]: E1123 23:24:15.379721 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bjff2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-b2nkd_calico-system(82559865-90d3-4a8e-83a3-002c0c14990f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:15.379937 containerd[1527]: time="2025-11-23T23:24:15.379884880Z" level=info msg="Container d78aa13cf041729ea78cbfa2779edb55bce23906c9284a7b67d1d243ff7e649e: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:24:15.380355 containerd[1527]: time="2025-11-23T23:24:15.380323875Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 23 23:24:15.401239 containerd[1527]: time="2025-11-23T23:24:15.401079945Z" level=info msg="CreateContainer within sandbox \"db8139bd549109dc53860b31d4167d4cf4fffa2840e3e52cac05e5406220771d\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d78aa13cf041729ea78cbfa2779edb55bce23906c9284a7b67d1d243ff7e649e\"" Nov 23 23:24:15.403124 containerd[1527]: time="2025-11-23T23:24:15.402930286Z" level=info msg="StartContainer for \"d78aa13cf041729ea78cbfa2779edb55bce23906c9284a7b67d1d243ff7e649e\"" Nov 23 23:24:15.403926 containerd[1527]: time="2025-11-23T23:24:15.403897756Z" level=info msg="connecting to shim d78aa13cf041729ea78cbfa2779edb55bce23906c9284a7b67d1d243ff7e649e" address="unix:///run/containerd/s/d2287fcb7a8840dd391174a409ffdb0ea1e6822abcf11c8fc47957ec535129bf" protocol=ttrpc version=3 Nov 23 23:24:15.430714 systemd[1]: Started cri-containerd-d78aa13cf041729ea78cbfa2779edb55bce23906c9284a7b67d1d243ff7e649e.scope - libcontainer container d78aa13cf041729ea78cbfa2779edb55bce23906c9284a7b67d1d243ff7e649e. Nov 23 23:24:15.466633 containerd[1527]: time="2025-11-23T23:24:15.466578719Z" level=info msg="StartContainer for \"d78aa13cf041729ea78cbfa2779edb55bce23906c9284a7b67d1d243ff7e649e\" returns successfully" Nov 23 23:24:15.588368 containerd[1527]: time="2025-11-23T23:24:15.588311922Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:15.589388 containerd[1527]: time="2025-11-23T23:24:15.589328752Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 23 23:24:15.589441 containerd[1527]: time="2025-11-23T23:24:15.589392751Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 23 23:24:15.589638 kubelet[2655]: E1123 23:24:15.589575 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:15.589638 kubelet[2655]: E1123 23:24:15.589635 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:15.589909 kubelet[2655]: E1123 23:24:15.589855 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-whk88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-846996596c-s6g44_calico-apiserver(927c8d49-e51e-462f-a42c-bb251141354f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:15.590138 containerd[1527]: time="2025-11-23T23:24:15.590054824Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 23 23:24:15.591178 kubelet[2655]: E1123 23:24:15.591132 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-846996596c-s6g44" podUID="927c8d49-e51e-462f-a42c-bb251141354f" Nov 23 23:24:15.797780 containerd[1527]: time="2025-11-23T23:24:15.797645235Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:15.798851 containerd[1527]: time="2025-11-23T23:24:15.798806824Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 23 23:24:15.798947 containerd[1527]: time="2025-11-23T23:24:15.798896423Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 23 23:24:15.799125 kubelet[2655]: E1123 23:24:15.799068 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 23 23:24:15.799169 kubelet[2655]: E1123 23:24:15.799139 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 23 23:24:15.799520 kubelet[2655]: E1123 23:24:15.799296 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bjff2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-b2nkd_calico-system(82559865-90d3-4a8e-83a3-002c0c14990f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:15.800904 kubelet[2655]: E1123 23:24:15.800853 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:24:15.882059 containerd[1527]: time="2025-11-23T23:24:15.881973899Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-clqmv,Uid:5b58e49f-cc99-463b-9194-304f2691d0a8,Namespace:kube-system,Attempt:0,}" Nov 23 23:24:15.882842 containerd[1527]: time="2025-11-23T23:24:15.882182216Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59795bc4c9-k44tv,Uid:17e162c7-08ce-4bc1-bbf1-49ed806bcede,Namespace:calico-apiserver,Attempt:0,}" Nov 23 23:24:16.017821 systemd-networkd[1446]: cali4c5aba257df: Link UP Nov 23 23:24:16.018853 systemd-networkd[1446]: cali4c5aba257df: Gained carrier Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.928 [INFO][4781] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0 calico-apiserver-59795bc4c9- calico-apiserver 17e162c7-08ce-4bc1-bbf1-49ed806bcede 807 0 2025-11-23 23:23:45 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:59795bc4c9 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-59795bc4c9-k44tv eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali4c5aba257df [] [] }} ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-k44tv" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.929 [INFO][4781] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-k44tv" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.971 [INFO][4803] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" HandleID="k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Workload="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.971 [INFO][4803] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" HandleID="k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Workload="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400012e570), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-59795bc4c9-k44tv", "timestamp":"2025-11-23 23:24:15.971213312 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.971 [INFO][4803] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.971 [INFO][4803] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.971 [INFO][4803] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.981 [INFO][4803] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.988 [INFO][4803] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.993 [INFO][4803] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.995 [INFO][4803] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.998 [INFO][4803] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.998 [INFO][4803] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:15.999 [INFO][4803] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:16.003 [INFO][4803] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:16.012 [INFO][4803] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:16.012 [INFO][4803] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" host="localhost" Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:16.012 [INFO][4803] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:16.033501 containerd[1527]: 2025-11-23 23:24:16.012 [INFO][4803] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" HandleID="k8s-pod-network.21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Workload="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" Nov 23 23:24:16.034048 containerd[1527]: 2025-11-23 23:24:16.014 [INFO][4781] cni-plugin/k8s.go 418: Populated endpoint ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-k44tv" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0", GenerateName:"calico-apiserver-59795bc4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"17e162c7-08ce-4bc1-bbf1-49ed806bcede", ResourceVersion:"807", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59795bc4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-59795bc4c9-k44tv", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4c5aba257df", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:16.034048 containerd[1527]: 2025-11-23 23:24:16.014 [INFO][4781] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-k44tv" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" Nov 23 23:24:16.034048 containerd[1527]: 2025-11-23 23:24:16.014 [INFO][4781] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4c5aba257df ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-k44tv" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" Nov 23 23:24:16.034048 containerd[1527]: 2025-11-23 23:24:16.020 [INFO][4781] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-k44tv" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" Nov 23 23:24:16.034048 containerd[1527]: 2025-11-23 23:24:16.021 [INFO][4781] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-k44tv" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0", GenerateName:"calico-apiserver-59795bc4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"17e162c7-08ce-4bc1-bbf1-49ed806bcede", ResourceVersion:"807", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59795bc4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d", Pod:"calico-apiserver-59795bc4c9-k44tv", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4c5aba257df", MAC:"56:a7:fa:2e:60:47", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:16.034048 containerd[1527]: 2025-11-23 23:24:16.029 [INFO][4781] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" Namespace="calico-apiserver" Pod="calico-apiserver-59795bc4c9-k44tv" WorkloadEndpoint="localhost-k8s-calico--apiserver--59795bc4c9--k44tv-eth0" Nov 23 23:24:16.066294 containerd[1527]: time="2025-11-23T23:24:16.066241722Z" level=info msg="connecting to shim 21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d" address="unix:///run/containerd/s/f642321b9ee5c646780f3c0f3def9dea6e2a5d186df13ae19f82fd2cf666ce42" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:16.075134 kubelet[2655]: E1123 23:24:16.075066 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:24:16.076553 kubelet[2655]: E1123 23:24:16.075956 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-846996596c-s6g44" podUID="927c8d49-e51e-462f-a42c-bb251141354f" Nov 23 23:24:16.106547 kubelet[2655]: I1123 23:24:16.106464 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-vwvns" podStartSLOduration=40.093973447 podStartE2EDuration="40.093973447s" podCreationTimestamp="2025-11-23 23:23:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 23:24:16.090295963 +0000 UTC m=+45.280194388" watchObservedRunningTime="2025-11-23 23:24:16.093973447 +0000 UTC m=+45.283871872" Nov 23 23:24:16.116784 systemd[1]: Started cri-containerd-21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d.scope - libcontainer container 21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d. Nov 23 23:24:16.148674 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:16.192982 systemd-networkd[1446]: cali41cb6ce42b1: Link UP Nov 23 23:24:16.193234 systemd-networkd[1446]: cali41cb6ce42b1: Gained carrier Nov 23 23:24:16.215694 containerd[1527]: time="2025-11-23T23:24:16.215650160Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59795bc4c9-k44tv,Uid:17e162c7-08ce-4bc1-bbf1-49ed806bcede,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"21d3d34f2686a80fa792f85329e02f2b9352f6120f7de4a3906993a4eed61d6d\"" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:15.941 [INFO][4775] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--clqmv-eth0 coredns-668d6bf9bc- kube-system 5b58e49f-cc99-463b-9194-304f2691d0a8 798 0 2025-11-23 23:23:36 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-clqmv eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali41cb6ce42b1 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Namespace="kube-system" Pod="coredns-668d6bf9bc-clqmv" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--clqmv-" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:15.942 [INFO][4775] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Namespace="kube-system" Pod="coredns-668d6bf9bc-clqmv" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:15.983 [INFO][4811] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" HandleID="k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Workload="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:15.984 [INFO][4811] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" HandleID="k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Workload="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400024b2e0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-clqmv", "timestamp":"2025-11-23 23:24:15.983875463 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:15.984 [INFO][4811] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.012 [INFO][4811] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.012 [INFO][4811] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.089 [INFO][4811] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.141 [INFO][4811] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.151 [INFO][4811] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.155 [INFO][4811] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.161 [INFO][4811] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.162 [INFO][4811] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.165 [INFO][4811] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513 Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.172 [INFO][4811] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.183 [INFO][4811] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.137/26] block=192.168.88.128/26 handle="k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.183 [INFO][4811] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.137/26] handle="k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" host="localhost" Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.183 [INFO][4811] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 23 23:24:16.217043 containerd[1527]: 2025-11-23 23:24:16.183 [INFO][4811] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.137/26] IPv6=[] ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" HandleID="k8s-pod-network.798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Workload="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" Nov 23 23:24:16.219057 containerd[1527]: 2025-11-23 23:24:16.187 [INFO][4775] cni-plugin/k8s.go 418: Populated endpoint ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Namespace="kube-system" Pod="coredns-668d6bf9bc-clqmv" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--clqmv-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"5b58e49f-cc99-463b-9194-304f2691d0a8", ResourceVersion:"798", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-clqmv", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.137/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali41cb6ce42b1", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:16.219057 containerd[1527]: 2025-11-23 23:24:16.187 [INFO][4775] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.137/32] ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Namespace="kube-system" Pod="coredns-668d6bf9bc-clqmv" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" Nov 23 23:24:16.219057 containerd[1527]: 2025-11-23 23:24:16.188 [INFO][4775] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali41cb6ce42b1 ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Namespace="kube-system" Pod="coredns-668d6bf9bc-clqmv" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" Nov 23 23:24:16.219057 containerd[1527]: 2025-11-23 23:24:16.191 [INFO][4775] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Namespace="kube-system" Pod="coredns-668d6bf9bc-clqmv" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" Nov 23 23:24:16.219057 containerd[1527]: 2025-11-23 23:24:16.195 [INFO][4775] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Namespace="kube-system" Pod="coredns-668d6bf9bc-clqmv" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--clqmv-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"5b58e49f-cc99-463b-9194-304f2691d0a8", ResourceVersion:"798", Generation:0, CreationTimestamp:time.Date(2025, time.November, 23, 23, 23, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513", Pod:"coredns-668d6bf9bc-clqmv", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.137/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali41cb6ce42b1", MAC:"3a:66:02:ce:0c:b9", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 23 23:24:16.219057 containerd[1527]: 2025-11-23 23:24:16.212 [INFO][4775] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" Namespace="kube-system" Pod="coredns-668d6bf9bc-clqmv" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--clqmv-eth0" Nov 23 23:24:16.219057 containerd[1527]: time="2025-11-23T23:24:16.217792499Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 23 23:24:16.227710 systemd-networkd[1446]: cali01286db5d83: Gained IPv6LL Nov 23 23:24:16.253944 containerd[1527]: time="2025-11-23T23:24:16.253837821Z" level=info msg="connecting to shim 798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513" address="unix:///run/containerd/s/62435fd69c09800e0953b43a3691fbbe409c6426e4eb63fc97deb7a03d2592d6" namespace=k8s.io protocol=ttrpc version=3 Nov 23 23:24:16.289932 systemd[1]: Started cri-containerd-798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513.scope - libcontainer container 798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513. Nov 23 23:24:16.312926 systemd-resolved[1357]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 23 23:24:16.353972 containerd[1527]: time="2025-11-23T23:24:16.353769470Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-clqmv,Uid:5b58e49f-cc99-463b-9194-304f2691d0a8,Namespace:kube-system,Attempt:0,} returns sandbox id \"798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513\"" Nov 23 23:24:16.360413 containerd[1527]: time="2025-11-23T23:24:16.359610452Z" level=info msg="CreateContainer within sandbox \"798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 23 23:24:16.368499 containerd[1527]: time="2025-11-23T23:24:16.368439564Z" level=info msg="Container a1d56768214f43c95d50776a9ab376315cf77edf80d814140bf3c75a7365b768: CDI devices from CRI Config.CDIDevices: []" Nov 23 23:24:16.374890 containerd[1527]: time="2025-11-23T23:24:16.374844260Z" level=info msg="CreateContainer within sandbox \"798fe1f60c667c59f88023845495588d2d28470929f82474b5fab2000d9b7513\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a1d56768214f43c95d50776a9ab376315cf77edf80d814140bf3c75a7365b768\"" Nov 23 23:24:16.375621 containerd[1527]: time="2025-11-23T23:24:16.375574213Z" level=info msg="StartContainer for \"a1d56768214f43c95d50776a9ab376315cf77edf80d814140bf3c75a7365b768\"" Nov 23 23:24:16.377149 containerd[1527]: time="2025-11-23T23:24:16.377110998Z" level=info msg="connecting to shim a1d56768214f43c95d50776a9ab376315cf77edf80d814140bf3c75a7365b768" address="unix:///run/containerd/s/62435fd69c09800e0953b43a3691fbbe409c6426e4eb63fc97deb7a03d2592d6" protocol=ttrpc version=3 Nov 23 23:24:16.397677 systemd[1]: Started cri-containerd-a1d56768214f43c95d50776a9ab376315cf77edf80d814140bf3c75a7365b768.scope - libcontainer container a1d56768214f43c95d50776a9ab376315cf77edf80d814140bf3c75a7365b768. Nov 23 23:24:16.428229 containerd[1527]: time="2025-11-23T23:24:16.428191251Z" level=info msg="StartContainer for \"a1d56768214f43c95d50776a9ab376315cf77edf80d814140bf3c75a7365b768\" returns successfully" Nov 23 23:24:16.446599 containerd[1527]: time="2025-11-23T23:24:16.446554669Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:16.447847 containerd[1527]: time="2025-11-23T23:24:16.447703978Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 23 23:24:16.447847 containerd[1527]: time="2025-11-23T23:24:16.447820937Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 23 23:24:16.448025 kubelet[2655]: E1123 23:24:16.447956 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:16.448025 kubelet[2655]: E1123 23:24:16.448018 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:16.448248 kubelet[2655]: E1123 23:24:16.448172 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lckd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59795bc4c9-k44tv_calico-apiserver(17e162c7-08ce-4bc1-bbf1-49ed806bcede): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:16.449405 kubelet[2655]: E1123 23:24:16.449365 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" podUID="17e162c7-08ce-4bc1-bbf1-49ed806bcede" Nov 23 23:24:16.611867 systemd-networkd[1446]: calic23cb12fb4b: Gained IPv6LL Nov 23 23:24:16.770421 systemd[1]: Started sshd@8-10.0.0.134:22-10.0.0.1:53222.service - OpenSSH per-connection server daemon (10.0.0.1:53222). Nov 23 23:24:16.851744 sshd[4971]: Accepted publickey for core from 10.0.0.1 port 53222 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:16.854322 sshd-session[4971]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:16.861899 systemd-logind[1513]: New session 9 of user core. Nov 23 23:24:16.881839 systemd[1]: Started session-9.scope - Session 9 of User core. Nov 23 23:24:17.060668 systemd-networkd[1446]: cali5a72d75a239: Gained IPv6LL Nov 23 23:24:17.065457 sshd[4974]: Connection closed by 10.0.0.1 port 53222 Nov 23 23:24:17.066065 sshd-session[4971]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:17.072232 systemd[1]: sshd@8-10.0.0.134:22-10.0.0.1:53222.service: Deactivated successfully. Nov 23 23:24:17.075580 systemd[1]: session-9.scope: Deactivated successfully. Nov 23 23:24:17.076922 systemd-logind[1513]: Session 9 logged out. Waiting for processes to exit. Nov 23 23:24:17.082262 kubelet[2655]: E1123 23:24:17.078353 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" podUID="17e162c7-08ce-4bc1-bbf1-49ed806bcede" Nov 23 23:24:17.087535 kubelet[2655]: E1123 23:24:17.083702 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-846996596c-s6g44" podUID="927c8d49-e51e-462f-a42c-bb251141354f" Nov 23 23:24:17.087535 kubelet[2655]: E1123 23:24:17.084223 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:24:17.088850 systemd-logind[1513]: Removed session 9. Nov 23 23:24:17.155853 kubelet[2655]: I1123 23:24:17.154698 2655 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-clqmv" podStartSLOduration=41.154677719 podStartE2EDuration="41.154677719s" podCreationTimestamp="2025-11-23 23:23:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 23:24:17.154130724 +0000 UTC m=+46.344029149" watchObservedRunningTime="2025-11-23 23:24:17.154677719 +0000 UTC m=+46.344576144" Nov 23 23:24:17.763691 systemd-networkd[1446]: cali41cb6ce42b1: Gained IPv6LL Nov 23 23:24:17.827651 systemd-networkd[1446]: cali4c5aba257df: Gained IPv6LL Nov 23 23:24:18.084081 kubelet[2655]: E1123 23:24:18.083820 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" podUID="17e162c7-08ce-4bc1-bbf1-49ed806bcede" Nov 23 23:24:21.883779 containerd[1527]: time="2025-11-23T23:24:21.883737272Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 23 23:24:22.076946 systemd[1]: Started sshd@9-10.0.0.134:22-10.0.0.1:35344.service - OpenSSH per-connection server daemon (10.0.0.1:35344). Nov 23 23:24:22.138336 sshd[5002]: Accepted publickey for core from 10.0.0.1 port 35344 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:22.139920 sshd-session[5002]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:22.142625 containerd[1527]: time="2025-11-23T23:24:22.142549383Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:22.143615 containerd[1527]: time="2025-11-23T23:24:22.143556735Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 23 23:24:22.143698 containerd[1527]: time="2025-11-23T23:24:22.143666934Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 23 23:24:22.143893 kubelet[2655]: E1123 23:24:22.143795 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 23 23:24:22.144889 kubelet[2655]: E1123 23:24:22.144029 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 23 23:24:22.144889 kubelet[2655]: E1123 23:24:22.144553 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:605c7f6a192d4d0899f32eea354f7105,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kgd29,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5b9c899964-zqqgk_calico-system(47273dbf-2dbb-4b82-b6aa-30c6db81392e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:22.148104 containerd[1527]: time="2025-11-23T23:24:22.147234942Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 23 23:24:22.147529 systemd-logind[1513]: New session 10 of user core. Nov 23 23:24:22.156709 systemd[1]: Started session-10.scope - Session 10 of User core. Nov 23 23:24:22.321229 sshd[5005]: Connection closed by 10.0.0.1 port 35344 Nov 23 23:24:22.321798 sshd-session[5002]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:22.338757 systemd[1]: sshd@9-10.0.0.134:22-10.0.0.1:35344.service: Deactivated successfully. Nov 23 23:24:22.341750 systemd[1]: session-10.scope: Deactivated successfully. Nov 23 23:24:22.342890 systemd-logind[1513]: Session 10 logged out. Waiting for processes to exit. Nov 23 23:24:22.345116 systemd[1]: Started sshd@10-10.0.0.134:22-10.0.0.1:35352.service - OpenSSH per-connection server daemon (10.0.0.1:35352). Nov 23 23:24:22.346996 systemd-logind[1513]: Removed session 10. Nov 23 23:24:22.363365 containerd[1527]: time="2025-11-23T23:24:22.363123929Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:22.364648 containerd[1527]: time="2025-11-23T23:24:22.364607436Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 23 23:24:22.364722 containerd[1527]: time="2025-11-23T23:24:22.364647276Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 23 23:24:22.364869 kubelet[2655]: E1123 23:24:22.364828 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 23 23:24:22.364955 kubelet[2655]: E1123 23:24:22.364881 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 23 23:24:22.365178 kubelet[2655]: E1123 23:24:22.364995 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kgd29,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5b9c899964-zqqgk_calico-system(47273dbf-2dbb-4b82-b6aa-30c6db81392e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:22.366515 kubelet[2655]: E1123 23:24:22.366175 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5b9c899964-zqqgk" podUID="47273dbf-2dbb-4b82-b6aa-30c6db81392e" Nov 23 23:24:22.411652 sshd[5022]: Accepted publickey for core from 10.0.0.1 port 35352 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:22.413717 sshd-session[5022]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:22.417909 systemd-logind[1513]: New session 11 of user core. Nov 23 23:24:22.428630 systemd[1]: Started session-11.scope - Session 11 of User core. Nov 23 23:24:22.615309 sshd[5026]: Connection closed by 10.0.0.1 port 35352 Nov 23 23:24:22.615676 sshd-session[5022]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:22.626229 systemd[1]: sshd@10-10.0.0.134:22-10.0.0.1:35352.service: Deactivated successfully. Nov 23 23:24:22.631323 systemd[1]: session-11.scope: Deactivated successfully. Nov 23 23:24:22.632981 systemd-logind[1513]: Session 11 logged out. Waiting for processes to exit. Nov 23 23:24:22.636866 systemd-logind[1513]: Removed session 11. Nov 23 23:24:22.643398 systemd[1]: Started sshd@11-10.0.0.134:22-10.0.0.1:35356.service - OpenSSH per-connection server daemon (10.0.0.1:35356). Nov 23 23:24:22.712967 sshd[5037]: Accepted publickey for core from 10.0.0.1 port 35356 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:22.714335 sshd-session[5037]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:22.718310 systemd-logind[1513]: New session 12 of user core. Nov 23 23:24:22.726647 systemd[1]: Started session-12.scope - Session 12 of User core. Nov 23 23:24:22.868004 sshd[5040]: Connection closed by 10.0.0.1 port 35356 Nov 23 23:24:22.868725 sshd-session[5037]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:22.872340 systemd-logind[1513]: Session 12 logged out. Waiting for processes to exit. Nov 23 23:24:22.873022 systemd[1]: sshd@11-10.0.0.134:22-10.0.0.1:35356.service: Deactivated successfully. Nov 23 23:24:22.875100 systemd[1]: session-12.scope: Deactivated successfully. Nov 23 23:24:22.877508 systemd-logind[1513]: Removed session 12. Nov 23 23:24:24.883618 containerd[1527]: time="2025-11-23T23:24:24.883272086Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 23 23:24:25.091058 containerd[1527]: time="2025-11-23T23:24:25.091013898Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:25.091964 containerd[1527]: time="2025-11-23T23:24:25.091931451Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 23 23:24:25.092048 containerd[1527]: time="2025-11-23T23:24:25.092008770Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 23 23:24:25.092228 kubelet[2655]: E1123 23:24:25.092167 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:25.092572 kubelet[2655]: E1123 23:24:25.092225 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:25.092572 kubelet[2655]: E1123 23:24:25.092356 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vr98w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59795bc4c9-sq9fp_calico-apiserver(d63258be-88c5-4ec0-93e4-199929ede838): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:25.093847 kubelet[2655]: E1123 23:24:25.093798 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" podUID="d63258be-88c5-4ec0-93e4-199929ede838" Nov 23 23:24:26.883014 containerd[1527]: time="2025-11-23T23:24:26.882761148Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 23 23:24:27.090072 containerd[1527]: time="2025-11-23T23:24:27.090007977Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:27.091007 containerd[1527]: time="2025-11-23T23:24:27.090972449Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 23 23:24:27.091062 containerd[1527]: time="2025-11-23T23:24:27.091047168Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 23 23:24:27.091211 kubelet[2655]: E1123 23:24:27.091170 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 23 23:24:27.091799 kubelet[2655]: E1123 23:24:27.091266 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 23 23:24:27.091799 kubelet[2655]: E1123 23:24:27.091496 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ddfs2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5c9f86c848-rx7m5_calico-system(31b181b6-1069-492f-a0d6-9078f4ef9b6a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:27.092627 containerd[1527]: time="2025-11-23T23:24:27.092295438Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 23 23:24:27.092962 kubelet[2655]: E1123 23:24:27.092916 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" podUID="31b181b6-1069-492f-a0d6-9078f4ef9b6a" Nov 23 23:24:27.299684 containerd[1527]: time="2025-11-23T23:24:27.299509721Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:27.302846 containerd[1527]: time="2025-11-23T23:24:27.302754454Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 23 23:24:27.302846 containerd[1527]: time="2025-11-23T23:24:27.302796894Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 23 23:24:27.303010 kubelet[2655]: E1123 23:24:27.302956 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 23 23:24:27.303057 kubelet[2655]: E1123 23:24:27.303024 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 23 23:24:27.303195 kubelet[2655]: E1123 23:24:27.303147 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-94mjb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-4h52k_calico-system(5c544408-aa7d-4999-b85e-a233becf4f04): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:27.304360 kubelet[2655]: E1123 23:24:27.304326 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-4h52k" podUID="5c544408-aa7d-4999-b85e-a233becf4f04" Nov 23 23:24:27.887897 systemd[1]: Started sshd@12-10.0.0.134:22-10.0.0.1:35372.service - OpenSSH per-connection server daemon (10.0.0.1:35372). Nov 23 23:24:27.948970 sshd[5056]: Accepted publickey for core from 10.0.0.1 port 35372 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:27.950243 sshd-session[5056]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:27.954256 systemd-logind[1513]: New session 13 of user core. Nov 23 23:24:27.960649 systemd[1]: Started session-13.scope - Session 13 of User core. Nov 23 23:24:28.084287 sshd[5059]: Connection closed by 10.0.0.1 port 35372 Nov 23 23:24:28.084234 sshd-session[5056]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:28.097608 systemd[1]: sshd@12-10.0.0.134:22-10.0.0.1:35372.service: Deactivated successfully. Nov 23 23:24:28.099115 systemd[1]: session-13.scope: Deactivated successfully. Nov 23 23:24:28.099756 systemd-logind[1513]: Session 13 logged out. Waiting for processes to exit. Nov 23 23:24:28.101951 systemd[1]: Started sshd@13-10.0.0.134:22-10.0.0.1:35384.service - OpenSSH per-connection server daemon (10.0.0.1:35384). Nov 23 23:24:28.102420 systemd-logind[1513]: Removed session 13. Nov 23 23:24:28.164037 sshd[5072]: Accepted publickey for core from 10.0.0.1 port 35384 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:28.165207 sshd-session[5072]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:28.169204 systemd-logind[1513]: New session 14 of user core. Nov 23 23:24:28.184620 systemd[1]: Started session-14.scope - Session 14 of User core. Nov 23 23:24:28.428043 sshd[5075]: Connection closed by 10.0.0.1 port 35384 Nov 23 23:24:28.428463 sshd-session[5072]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:28.436595 systemd[1]: sshd@13-10.0.0.134:22-10.0.0.1:35384.service: Deactivated successfully. Nov 23 23:24:28.438459 systemd[1]: session-14.scope: Deactivated successfully. Nov 23 23:24:28.439309 systemd-logind[1513]: Session 14 logged out. Waiting for processes to exit. Nov 23 23:24:28.442277 systemd[1]: Started sshd@14-10.0.0.134:22-10.0.0.1:35392.service - OpenSSH per-connection server daemon (10.0.0.1:35392). Nov 23 23:24:28.443069 systemd-logind[1513]: Removed session 14. Nov 23 23:24:28.499204 sshd[5087]: Accepted publickey for core from 10.0.0.1 port 35392 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:28.500380 sshd-session[5087]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:28.504534 systemd-logind[1513]: New session 15 of user core. Nov 23 23:24:28.513633 systemd[1]: Started session-15.scope - Session 15 of User core. Nov 23 23:24:28.883219 containerd[1527]: time="2025-11-23T23:24:28.883107198Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 23 23:24:29.097201 containerd[1527]: time="2025-11-23T23:24:29.097016580Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:29.098952 containerd[1527]: time="2025-11-23T23:24:29.098528288Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 23 23:24:29.098952 containerd[1527]: time="2025-11-23T23:24:29.098567928Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 23 23:24:29.099070 kubelet[2655]: E1123 23:24:29.098854 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 23 23:24:29.099070 kubelet[2655]: E1123 23:24:29.099023 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 23 23:24:29.099453 kubelet[2655]: E1123 23:24:29.099284 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bjff2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-b2nkd_calico-system(82559865-90d3-4a8e-83a3-002c0c14990f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:29.102137 containerd[1527]: time="2025-11-23T23:24:29.101902022Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 23 23:24:29.127287 sshd[5090]: Connection closed by 10.0.0.1 port 35392 Nov 23 23:24:29.127933 sshd-session[5087]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:29.140551 systemd[1]: sshd@14-10.0.0.134:22-10.0.0.1:35392.service: Deactivated successfully. Nov 23 23:24:29.142941 systemd[1]: session-15.scope: Deactivated successfully. Nov 23 23:24:29.144139 systemd-logind[1513]: Session 15 logged out. Waiting for processes to exit. Nov 23 23:24:29.148329 systemd[1]: Started sshd@15-10.0.0.134:22-10.0.0.1:35398.service - OpenSSH per-connection server daemon (10.0.0.1:35398). Nov 23 23:24:29.149894 systemd-logind[1513]: Removed session 15. Nov 23 23:24:29.215427 sshd[5110]: Accepted publickey for core from 10.0.0.1 port 35398 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:29.216609 sshd-session[5110]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:29.220844 systemd-logind[1513]: New session 16 of user core. Nov 23 23:24:29.231679 systemd[1]: Started session-16.scope - Session 16 of User core. Nov 23 23:24:29.308818 containerd[1527]: time="2025-11-23T23:24:29.308607433Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:29.309657 containerd[1527]: time="2025-11-23T23:24:29.309556465Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 23 23:24:29.309657 containerd[1527]: time="2025-11-23T23:24:29.309626865Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 23 23:24:29.309836 kubelet[2655]: E1123 23:24:29.309764 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 23 23:24:29.309836 kubelet[2655]: E1123 23:24:29.309822 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 23 23:24:29.310012 kubelet[2655]: E1123 23:24:29.309936 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bjff2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-b2nkd_calico-system(82559865-90d3-4a8e-83a3-002c0c14990f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:29.311370 kubelet[2655]: E1123 23:24:29.311322 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:24:29.527374 sshd[5114]: Connection closed by 10.0.0.1 port 35398 Nov 23 23:24:29.529571 sshd-session[5110]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:29.537862 systemd[1]: sshd@15-10.0.0.134:22-10.0.0.1:35398.service: Deactivated successfully. Nov 23 23:24:29.540569 systemd[1]: session-16.scope: Deactivated successfully. Nov 23 23:24:29.544221 systemd-logind[1513]: Session 16 logged out. Waiting for processes to exit. Nov 23 23:24:29.547767 systemd[1]: Started sshd@16-10.0.0.134:22-10.0.0.1:54434.service - OpenSSH per-connection server daemon (10.0.0.1:54434). Nov 23 23:24:29.550316 systemd-logind[1513]: Removed session 16. Nov 23 23:24:29.607366 sshd[5125]: Accepted publickey for core from 10.0.0.1 port 54434 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:29.609088 sshd-session[5125]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:29.613534 systemd-logind[1513]: New session 17 of user core. Nov 23 23:24:29.623651 systemd[1]: Started session-17.scope - Session 17 of User core. Nov 23 23:24:29.772511 sshd[5128]: Connection closed by 10.0.0.1 port 54434 Nov 23 23:24:29.772623 sshd-session[5125]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:29.776012 systemd[1]: sshd@16-10.0.0.134:22-10.0.0.1:54434.service: Deactivated successfully. Nov 23 23:24:29.778059 systemd[1]: session-17.scope: Deactivated successfully. Nov 23 23:24:29.779695 systemd-logind[1513]: Session 17 logged out. Waiting for processes to exit. Nov 23 23:24:29.780986 systemd-logind[1513]: Removed session 17. Nov 23 23:24:30.884300 containerd[1527]: time="2025-11-23T23:24:30.884060062Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 23 23:24:31.098096 containerd[1527]: time="2025-11-23T23:24:31.097997605Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:31.099338 containerd[1527]: time="2025-11-23T23:24:31.099264315Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 23 23:24:31.099430 containerd[1527]: time="2025-11-23T23:24:31.099372315Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 23 23:24:31.099728 kubelet[2655]: E1123 23:24:31.099677 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:31.100817 kubelet[2655]: E1123 23:24:31.099730 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:31.100817 kubelet[2655]: E1123 23:24:31.099890 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-whk88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-846996596c-s6g44_calico-apiserver(927c8d49-e51e-462f-a42c-bb251141354f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:31.103123 kubelet[2655]: E1123 23:24:31.102955 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-846996596c-s6g44" podUID="927c8d49-e51e-462f-a42c-bb251141354f" Nov 23 23:24:32.882402 containerd[1527]: time="2025-11-23T23:24:32.882356033Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 23 23:24:33.096427 containerd[1527]: time="2025-11-23T23:24:33.096363132Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Nov 23 23:24:33.097395 containerd[1527]: time="2025-11-23T23:24:33.097352844Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 23 23:24:33.097477 containerd[1527]: time="2025-11-23T23:24:33.097436764Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 23 23:24:33.097667 kubelet[2655]: E1123 23:24:33.097604 2655 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:33.097953 kubelet[2655]: E1123 23:24:33.097667 2655 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 23 23:24:33.097953 kubelet[2655]: E1123 23:24:33.097797 2655 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lckd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-59795bc4c9-k44tv_calico-apiserver(17e162c7-08ce-4bc1-bbf1-49ed806bcede): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 23 23:24:33.099045 kubelet[2655]: E1123 23:24:33.099012 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" podUID="17e162c7-08ce-4bc1-bbf1-49ed806bcede" Nov 23 23:24:34.783871 systemd[1]: Started sshd@17-10.0.0.134:22-10.0.0.1:54444.service - OpenSSH per-connection server daemon (10.0.0.1:54444). Nov 23 23:24:34.848596 sshd[5151]: Accepted publickey for core from 10.0.0.1 port 54444 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:34.850009 sshd-session[5151]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:34.854897 systemd-logind[1513]: New session 18 of user core. Nov 23 23:24:34.862660 systemd[1]: Started session-18.scope - Session 18 of User core. Nov 23 23:24:34.994022 sshd[5154]: Connection closed by 10.0.0.1 port 54444 Nov 23 23:24:34.994584 sshd-session[5151]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:34.998356 systemd[1]: sshd@17-10.0.0.134:22-10.0.0.1:54444.service: Deactivated successfully. Nov 23 23:24:35.000307 systemd[1]: session-18.scope: Deactivated successfully. Nov 23 23:24:35.002386 systemd-logind[1513]: Session 18 logged out. Waiting for processes to exit. Nov 23 23:24:35.004040 systemd-logind[1513]: Removed session 18. Nov 23 23:24:37.882973 kubelet[2655]: E1123 23:24:37.882851 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5b9c899964-zqqgk" podUID="47273dbf-2dbb-4b82-b6aa-30c6db81392e" Nov 23 23:24:39.882929 kubelet[2655]: E1123 23:24:39.882876 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-sq9fp" podUID="d63258be-88c5-4ec0-93e4-199929ede838" Nov 23 23:24:40.008855 systemd[1]: Started sshd@18-10.0.0.134:22-10.0.0.1:43352.service - OpenSSH per-connection server daemon (10.0.0.1:43352). Nov 23 23:24:40.067300 sshd[5197]: Accepted publickey for core from 10.0.0.1 port 43352 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:40.068974 sshd-session[5197]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:40.077061 systemd-logind[1513]: New session 19 of user core. Nov 23 23:24:40.085629 systemd[1]: Started session-19.scope - Session 19 of User core. Nov 23 23:24:40.236035 sshd[5200]: Connection closed by 10.0.0.1 port 43352 Nov 23 23:24:40.235818 sshd-session[5197]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:40.240200 systemd[1]: sshd@18-10.0.0.134:22-10.0.0.1:43352.service: Deactivated successfully. Nov 23 23:24:40.242290 systemd[1]: session-19.scope: Deactivated successfully. Nov 23 23:24:40.246476 systemd-logind[1513]: Session 19 logged out. Waiting for processes to exit. Nov 23 23:24:40.248614 systemd-logind[1513]: Removed session 19. Nov 23 23:24:40.882872 kubelet[2655]: E1123 23:24:40.882808 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-4h52k" podUID="5c544408-aa7d-4999-b85e-a233becf4f04" Nov 23 23:24:41.883249 kubelet[2655]: E1123 23:24:41.883172 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5c9f86c848-rx7m5" podUID="31b181b6-1069-492f-a0d6-9078f4ef9b6a" Nov 23 23:24:41.884323 kubelet[2655]: E1123 23:24:41.883952 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-b2nkd" podUID="82559865-90d3-4a8e-83a3-002c0c14990f" Nov 23 23:24:44.882886 kubelet[2655]: E1123 23:24:44.882685 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-846996596c-s6g44" podUID="927c8d49-e51e-462f-a42c-bb251141354f" Nov 23 23:24:45.247728 systemd[1]: Started sshd@19-10.0.0.134:22-10.0.0.1:43362.service - OpenSSH per-connection server daemon (10.0.0.1:43362). Nov 23 23:24:45.292546 sshd[5216]: Accepted publickey for core from 10.0.0.1 port 43362 ssh2: RSA SHA256:xK0odXIrRLy2uvFTHd2XiQ92YaTCLtqdWVOOXxQURNk Nov 23 23:24:45.293414 sshd-session[5216]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 23 23:24:45.298431 systemd-logind[1513]: New session 20 of user core. Nov 23 23:24:45.301696 systemd[1]: Started session-20.scope - Session 20 of User core. Nov 23 23:24:45.425381 sshd[5220]: Connection closed by 10.0.0.1 port 43362 Nov 23 23:24:45.425715 sshd-session[5216]: pam_unix(sshd:session): session closed for user core Nov 23 23:24:45.429398 systemd[1]: sshd@19-10.0.0.134:22-10.0.0.1:43362.service: Deactivated successfully. Nov 23 23:24:45.431125 systemd[1]: session-20.scope: Deactivated successfully. Nov 23 23:24:45.431761 systemd-logind[1513]: Session 20 logged out. Waiting for processes to exit. Nov 23 23:24:45.432677 systemd-logind[1513]: Removed session 20. Nov 23 23:24:46.885543 kubelet[2655]: E1123 23:24:46.885462 2655 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-59795bc4c9-k44tv" podUID="17e162c7-08ce-4bc1-bbf1-49ed806bcede"